var/home/core/zuul-output/0000755000175000017500000000000015116340516014527 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015116355527015503 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005303622615116355520017707 0ustar rootrootDec 10 18:55:18 crc systemd[1]: Starting Kubernetes Kubelet... Dec 10 18:55:18 crc restorecon[4701]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:18 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:19 crc restorecon[4701]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 10 18:55:19 crc restorecon[4701]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 10 18:55:19 crc kubenswrapper[4828]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 18:55:19 crc kubenswrapper[4828]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 10 18:55:19 crc kubenswrapper[4828]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 18:55:19 crc kubenswrapper[4828]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 18:55:19 crc kubenswrapper[4828]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 10 18:55:19 crc kubenswrapper[4828]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.618980 4828 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622883 4828 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622900 4828 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622905 4828 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622909 4828 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622913 4828 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622917 4828 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622922 4828 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622926 4828 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622930 4828 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622935 4828 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622940 4828 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622945 4828 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622952 4828 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622959 4828 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622965 4828 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622969 4828 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622974 4828 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622979 4828 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622983 4828 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622987 4828 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622991 4828 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622996 4828 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.622999 4828 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623003 4828 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623007 4828 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623010 4828 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623014 4828 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623017 4828 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623021 4828 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623025 4828 feature_gate.go:330] unrecognized feature gate: Example Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623029 4828 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623032 4828 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623036 4828 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623039 4828 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623043 4828 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623046 4828 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623050 4828 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623054 4828 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623057 4828 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623061 4828 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623064 4828 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623068 4828 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623071 4828 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623076 4828 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623080 4828 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623085 4828 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623089 4828 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623092 4828 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623097 4828 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623101 4828 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623104 4828 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623107 4828 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623111 4828 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623114 4828 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623118 4828 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623121 4828 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623125 4828 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623128 4828 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623133 4828 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623138 4828 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623142 4828 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623146 4828 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623149 4828 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623153 4828 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623157 4828 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623160 4828 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623164 4828 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623167 4828 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623171 4828 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623174 4828 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.623177 4828 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623699 4828 flags.go:64] FLAG: --address="0.0.0.0" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623714 4828 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623724 4828 flags.go:64] FLAG: --anonymous-auth="true" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623730 4828 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623737 4828 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623742 4828 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623750 4828 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623757 4828 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623762 4828 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623767 4828 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623772 4828 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623778 4828 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623783 4828 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623788 4828 flags.go:64] FLAG: --cgroup-root="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623793 4828 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623815 4828 flags.go:64] FLAG: --client-ca-file="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623820 4828 flags.go:64] FLAG: --cloud-config="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623825 4828 flags.go:64] FLAG: --cloud-provider="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623830 4828 flags.go:64] FLAG: --cluster-dns="[]" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623837 4828 flags.go:64] FLAG: --cluster-domain="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623843 4828 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623848 4828 flags.go:64] FLAG: --config-dir="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623853 4828 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623859 4828 flags.go:64] FLAG: --container-log-max-files="5" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623867 4828 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623872 4828 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623878 4828 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623883 4828 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623888 4828 flags.go:64] FLAG: --contention-profiling="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623893 4828 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623897 4828 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623902 4828 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623907 4828 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623914 4828 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623919 4828 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623923 4828 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623928 4828 flags.go:64] FLAG: --enable-load-reader="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623933 4828 flags.go:64] FLAG: --enable-server="true" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623940 4828 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623947 4828 flags.go:64] FLAG: --event-burst="100" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623953 4828 flags.go:64] FLAG: --event-qps="50" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623958 4828 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623964 4828 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623969 4828 flags.go:64] FLAG: --eviction-hard="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623976 4828 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623982 4828 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623987 4828 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623992 4828 flags.go:64] FLAG: --eviction-soft="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.623997 4828 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624002 4828 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624007 4828 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624012 4828 flags.go:64] FLAG: --experimental-mounter-path="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624017 4828 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624023 4828 flags.go:64] FLAG: --fail-swap-on="true" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624028 4828 flags.go:64] FLAG: --feature-gates="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624033 4828 flags.go:64] FLAG: --file-check-frequency="20s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624037 4828 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624042 4828 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624046 4828 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624051 4828 flags.go:64] FLAG: --healthz-port="10248" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624055 4828 flags.go:64] FLAG: --help="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624059 4828 flags.go:64] FLAG: --hostname-override="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624063 4828 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624067 4828 flags.go:64] FLAG: --http-check-frequency="20s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624071 4828 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624075 4828 flags.go:64] FLAG: --image-credential-provider-config="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624079 4828 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624083 4828 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624087 4828 flags.go:64] FLAG: --image-service-endpoint="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624091 4828 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624097 4828 flags.go:64] FLAG: --kube-api-burst="100" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624101 4828 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624105 4828 flags.go:64] FLAG: --kube-api-qps="50" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624109 4828 flags.go:64] FLAG: --kube-reserved="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624113 4828 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624117 4828 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624121 4828 flags.go:64] FLAG: --kubelet-cgroups="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624125 4828 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624129 4828 flags.go:64] FLAG: --lock-file="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624133 4828 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624137 4828 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624140 4828 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624147 4828 flags.go:64] FLAG: --log-json-split-stream="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624151 4828 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624155 4828 flags.go:64] FLAG: --log-text-split-stream="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624160 4828 flags.go:64] FLAG: --logging-format="text" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624164 4828 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624168 4828 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624172 4828 flags.go:64] FLAG: --manifest-url="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624176 4828 flags.go:64] FLAG: --manifest-url-header="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624183 4828 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624187 4828 flags.go:64] FLAG: --max-open-files="1000000" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624192 4828 flags.go:64] FLAG: --max-pods="110" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624196 4828 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624200 4828 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624204 4828 flags.go:64] FLAG: --memory-manager-policy="None" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624208 4828 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624212 4828 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624216 4828 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624220 4828 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624231 4828 flags.go:64] FLAG: --node-status-max-images="50" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624235 4828 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624239 4828 flags.go:64] FLAG: --oom-score-adj="-999" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624243 4828 flags.go:64] FLAG: --pod-cidr="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624247 4828 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624253 4828 flags.go:64] FLAG: --pod-manifest-path="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624257 4828 flags.go:64] FLAG: --pod-max-pids="-1" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624261 4828 flags.go:64] FLAG: --pods-per-core="0" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624265 4828 flags.go:64] FLAG: --port="10250" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624269 4828 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624273 4828 flags.go:64] FLAG: --provider-id="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624277 4828 flags.go:64] FLAG: --qos-reserved="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624281 4828 flags.go:64] FLAG: --read-only-port="10255" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624285 4828 flags.go:64] FLAG: --register-node="true" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624289 4828 flags.go:64] FLAG: --register-schedulable="true" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624293 4828 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624300 4828 flags.go:64] FLAG: --registry-burst="10" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624303 4828 flags.go:64] FLAG: --registry-qps="5" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624307 4828 flags.go:64] FLAG: --reserved-cpus="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624311 4828 flags.go:64] FLAG: --reserved-memory="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624317 4828 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624322 4828 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624326 4828 flags.go:64] FLAG: --rotate-certificates="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624330 4828 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624334 4828 flags.go:64] FLAG: --runonce="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624338 4828 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624342 4828 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624346 4828 flags.go:64] FLAG: --seccomp-default="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624350 4828 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624354 4828 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624358 4828 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624362 4828 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624366 4828 flags.go:64] FLAG: --storage-driver-password="root" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624370 4828 flags.go:64] FLAG: --storage-driver-secure="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624374 4828 flags.go:64] FLAG: --storage-driver-table="stats" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624378 4828 flags.go:64] FLAG: --storage-driver-user="root" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624382 4828 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624386 4828 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624390 4828 flags.go:64] FLAG: --system-cgroups="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624394 4828 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624400 4828 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624404 4828 flags.go:64] FLAG: --tls-cert-file="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624408 4828 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624413 4828 flags.go:64] FLAG: --tls-min-version="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624416 4828 flags.go:64] FLAG: --tls-private-key-file="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624420 4828 flags.go:64] FLAG: --topology-manager-policy="none" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624424 4828 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624428 4828 flags.go:64] FLAG: --topology-manager-scope="container" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624432 4828 flags.go:64] FLAG: --v="2" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624437 4828 flags.go:64] FLAG: --version="false" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624443 4828 flags.go:64] FLAG: --vmodule="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624453 4828 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624457 4828 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624548 4828 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624553 4828 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624557 4828 feature_gate.go:330] unrecognized feature gate: Example Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624561 4828 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624566 4828 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624570 4828 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624574 4828 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624578 4828 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624582 4828 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624585 4828 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624589 4828 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624593 4828 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624596 4828 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624599 4828 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624603 4828 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624607 4828 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624610 4828 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624614 4828 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624617 4828 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624621 4828 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624624 4828 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624628 4828 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624635 4828 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624638 4828 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624642 4828 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624646 4828 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624649 4828 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624652 4828 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624656 4828 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624659 4828 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624663 4828 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624666 4828 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624669 4828 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624673 4828 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624676 4828 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624680 4828 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624683 4828 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624688 4828 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624692 4828 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624696 4828 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624705 4828 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624709 4828 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624713 4828 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624716 4828 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624720 4828 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624724 4828 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624728 4828 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624731 4828 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624735 4828 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624738 4828 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624743 4828 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624747 4828 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624751 4828 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624754 4828 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624760 4828 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624764 4828 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624767 4828 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624772 4828 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624777 4828 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624780 4828 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624784 4828 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624789 4828 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624793 4828 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624813 4828 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624817 4828 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624820 4828 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624825 4828 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624829 4828 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624833 4828 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624837 4828 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.624841 4828 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.624847 4828 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.637090 4828 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.637137 4828 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637291 4828 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637309 4828 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637320 4828 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637331 4828 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637342 4828 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637353 4828 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637367 4828 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637379 4828 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637389 4828 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637400 4828 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637411 4828 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637426 4828 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637439 4828 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637449 4828 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637460 4828 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637470 4828 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637480 4828 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637491 4828 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637500 4828 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637511 4828 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637521 4828 feature_gate.go:330] unrecognized feature gate: Example Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637531 4828 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637541 4828 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637551 4828 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637561 4828 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637571 4828 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637581 4828 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637591 4828 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637600 4828 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637610 4828 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637621 4828 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637633 4828 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637646 4828 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637659 4828 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637673 4828 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637684 4828 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637694 4828 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637704 4828 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637714 4828 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637724 4828 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637734 4828 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637744 4828 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637754 4828 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637764 4828 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637777 4828 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637789 4828 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637861 4828 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637873 4828 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637883 4828 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637896 4828 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637906 4828 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637919 4828 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637930 4828 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637942 4828 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637952 4828 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637962 4828 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637972 4828 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637982 4828 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.637992 4828 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638003 4828 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638013 4828 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638024 4828 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638034 4828 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638046 4828 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638056 4828 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638066 4828 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638077 4828 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638087 4828 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638096 4828 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638106 4828 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638119 4828 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.638136 4828 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638431 4828 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638451 4828 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638462 4828 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638473 4828 feature_gate.go:330] unrecognized feature gate: Example Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638485 4828 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638499 4828 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638515 4828 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638526 4828 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638537 4828 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638549 4828 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638560 4828 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638572 4828 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638584 4828 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638595 4828 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638605 4828 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638616 4828 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638626 4828 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638636 4828 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638646 4828 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638656 4828 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638668 4828 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638678 4828 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638688 4828 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638699 4828 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638709 4828 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638719 4828 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638729 4828 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638739 4828 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638749 4828 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638759 4828 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638770 4828 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638780 4828 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638790 4828 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638834 4828 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638849 4828 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638860 4828 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638871 4828 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638881 4828 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638891 4828 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638901 4828 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638911 4828 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638922 4828 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638933 4828 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638944 4828 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638954 4828 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638963 4828 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638973 4828 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638984 4828 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.638993 4828 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639003 4828 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639017 4828 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639030 4828 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639041 4828 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639051 4828 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639061 4828 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639070 4828 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639081 4828 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639091 4828 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639101 4828 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639111 4828 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639120 4828 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639130 4828 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639144 4828 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639156 4828 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639167 4828 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639177 4828 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639187 4828 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639197 4828 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639206 4828 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639216 4828 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.639227 4828 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.639243 4828 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.639968 4828 server.go:940] "Client rotation is on, will bootstrap in background" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.646680 4828 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.646854 4828 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.647738 4828 server.go:997] "Starting client certificate rotation" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.647777 4828 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.647937 4828 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-01 11:42:14.836489243 +0000 UTC Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.648032 4828 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.656063 4828 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 10 18:55:19 crc kubenswrapper[4828]: E1210 18:55:19.658053 4828 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.659424 4828 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.667776 4828 log.go:25] "Validated CRI v1 runtime API" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.681729 4828 log.go:25] "Validated CRI v1 image API" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.683029 4828 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.685626 4828 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-10-18-50-55-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.685677 4828 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.712139 4828 manager.go:217] Machine: {Timestamp:2025-12-10 18:55:19.709908545 +0000 UTC m=+0.220519620 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654132736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:2c4a1f25-f3cd-4108-b88c-83e56814cf09 BootID:df53ff5c-912c-4420-b6db-884332bb2d58 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730829824 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827068416 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:6b:b5:a3 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:6b:b5:a3 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:de:6a:5c Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:bc:22:67 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:d6:e4:92 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:b5:99:f8 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:8e:af:cc:9e:15:12 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:9a:e6:7c:13:b9:9f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654132736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.712583 4828 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.712852 4828 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.713926 4828 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.714247 4828 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.714310 4828 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.714834 4828 topology_manager.go:138] "Creating topology manager with none policy" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.714855 4828 container_manager_linux.go:303] "Creating device plugin manager" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.715176 4828 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.715219 4828 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.715582 4828 state_mem.go:36] "Initialized new in-memory state store" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.716787 4828 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.718435 4828 kubelet.go:418] "Attempting to sync node with API server" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.718477 4828 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.718523 4828 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.718546 4828 kubelet.go:324] "Adding apiserver pod source" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.718564 4828 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.721025 4828 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.721283 4828 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.721412 4828 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.721431 4828 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Dec 10 18:55:19 crc kubenswrapper[4828]: E1210 18:55:19.721484 4828 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Dec 10 18:55:19 crc kubenswrapper[4828]: E1210 18:55:19.721543 4828 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.722372 4828 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.723037 4828 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.723069 4828 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.723081 4828 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.723092 4828 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.723107 4828 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.723116 4828 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.723126 4828 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.723144 4828 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.723155 4828 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.723165 4828 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.723178 4828 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.723187 4828 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.723410 4828 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.724051 4828 server.go:1280] "Started kubelet" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.724337 4828 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.724340 4828 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.724952 4828 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 10 18:55:19 crc systemd[1]: Started Kubernetes Kubelet. Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.725851 4828 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.726777 4828 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 10 18:55:19 crc kubenswrapper[4828]: E1210 18:55:19.726246 4828 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.214:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187fef81cd23c1fa default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 18:55:19.724016122 +0000 UTC m=+0.234627137,LastTimestamp:2025-12-10 18:55:19.724016122 +0000 UTC m=+0.234627137,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.726863 4828 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.726937 4828 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 23:50:08.977445331 +0000 UTC Dec 10 18:55:19 crc kubenswrapper[4828]: E1210 18:55:19.727511 4828 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.727573 4828 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.727584 4828 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.727716 4828 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.728615 4828 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Dec 10 18:55:19 crc kubenswrapper[4828]: E1210 18:55:19.728704 4828 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.728849 4828 factory.go:55] Registering systemd factory Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.728879 4828 factory.go:221] Registration of the systemd container factory successfully Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.729238 4828 factory.go:153] Registering CRI-O factory Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.729264 4828 factory.go:221] Registration of the crio container factory successfully Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.729450 4828 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.729492 4828 factory.go:103] Registering Raw factory Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.729514 4828 manager.go:1196] Started watching for new ooms in manager Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.729695 4828 server.go:460] "Adding debug handlers to kubelet server" Dec 10 18:55:19 crc kubenswrapper[4828]: E1210 18:55:19.729750 4828 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="200ms" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.730592 4828 manager.go:319] Starting recovery of all containers Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748047 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748180 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748209 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748228 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748246 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748266 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748285 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748304 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748326 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748343 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748361 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748383 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748403 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748424 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748443 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748460 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748479 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748496 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748513 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748530 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748548 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748568 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748586 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748603 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748654 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748673 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748695 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748716 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748735 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748754 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748772 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748791 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748841 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748922 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748942 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748961 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.748981 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749005 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749025 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749043 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749062 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749082 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749100 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749122 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749140 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749160 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749179 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749197 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749219 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749246 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749272 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749299 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749332 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749354 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749377 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749396 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749415 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749470 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749490 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749510 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749528 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749551 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749569 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749588 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749606 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749624 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749642 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749660 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749678 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749697 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749716 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749733 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749782 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749851 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749871 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749892 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749913 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749933 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749952 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.749972 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750001 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750020 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750041 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750098 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750119 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750139 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750157 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750176 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750195 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750214 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750232 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750255 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750274 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750292 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750309 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750326 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750350 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750376 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750404 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750429 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750452 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750476 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750499 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.750528 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751496 4828 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751570 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751596 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751621 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751643 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751663 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751684 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751706 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751728 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751749 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751770 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751788 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751842 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751862 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751879 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751900 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751921 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751939 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751958 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751978 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.751994 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752011 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752028 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752045 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752063 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752081 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752099 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752118 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752136 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752154 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752171 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752191 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752208 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752226 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752244 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752265 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752284 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752300 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752319 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752336 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752353 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752371 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752389 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752406 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752425 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752446 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752463 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752480 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752499 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752518 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752534 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752554 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752571 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752589 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752608 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752625 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752642 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752660 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752677 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752695 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752712 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752732 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752749 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752768 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752787 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752838 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752856 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752876 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752895 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752912 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752929 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752949 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752967 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.752985 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753003 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753028 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753051 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753076 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753102 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753128 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753151 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753173 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753192 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753212 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753230 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753249 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753266 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753283 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753302 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753327 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753350 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753373 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753396 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753420 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753438 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753457 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753474 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753493 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753511 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753533 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753554 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753570 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753588 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753605 4828 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753622 4828 reconstruct.go:97] "Volume reconstruction finished" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.753635 4828 reconciler.go:26] "Reconciler: start to sync state" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.754523 4828 manager.go:324] Recovery completed Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.766084 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.768044 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.768091 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.768104 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.769441 4828 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.769504 4828 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.769536 4828 state_mem.go:36] "Initialized new in-memory state store" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.782212 4828 policy_none.go:49] "None policy: Start" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.783196 4828 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.783235 4828 state_mem.go:35] "Initializing new in-memory state store" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.785482 4828 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.787385 4828 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.787441 4828 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.787481 4828 kubelet.go:2335] "Starting kubelet main sync loop" Dec 10 18:55:19 crc kubenswrapper[4828]: E1210 18:55:19.787546 4828 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 10 18:55:19 crc kubenswrapper[4828]: W1210 18:55:19.788142 4828 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Dec 10 18:55:19 crc kubenswrapper[4828]: E1210 18:55:19.788201 4828 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Dec 10 18:55:19 crc kubenswrapper[4828]: E1210 18:55:19.828567 4828 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.838339 4828 manager.go:334] "Starting Device Plugin manager" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.838425 4828 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.838450 4828 server.go:79] "Starting device plugin registration server" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.839133 4828 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.839164 4828 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.839301 4828 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.839381 4828 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.839388 4828 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 10 18:55:19 crc kubenswrapper[4828]: E1210 18:55:19.846571 4828 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.887660 4828 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.887740 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.888614 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.888667 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.888678 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.888830 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.888996 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.889028 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.889529 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.889814 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.889836 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.889854 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.889944 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.889976 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.890641 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.891009 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.891102 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.891443 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.891462 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.891476 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.891614 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.892077 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.892115 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.892715 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.892736 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.892748 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.892910 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.892925 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.892937 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.893017 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.893028 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.893038 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.893293 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.893327 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.893293 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.894743 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.894767 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.894775 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.894895 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.894915 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.895043 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.895076 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.895092 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.895396 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.895420 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.895428 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4828]: E1210 18:55:19.931055 4828 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="400ms" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.940139 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.941322 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.941348 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.941356 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.941374 4828 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 18:55:19 crc kubenswrapper[4828]: E1210 18:55:19.941858 4828 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.214:6443: connect: connection refused" node="crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.955866 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.955917 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.955953 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.956033 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.956094 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.956183 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.956213 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.956271 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.956323 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.956361 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.956391 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.956455 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.956606 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.956699 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:19 crc kubenswrapper[4828]: I1210 18:55:19.956735 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058480 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058529 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058556 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058579 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058605 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058625 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058644 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058665 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058667 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058690 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058710 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058732 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058770 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058827 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058795 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058827 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058869 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058891 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058873 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058900 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058702 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058932 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058940 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058965 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058922 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.058901 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.059000 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.059054 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.059110 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.059216 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.141987 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.143227 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.143284 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.143300 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.143322 4828 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 18:55:20 crc kubenswrapper[4828]: E1210 18:55:20.143833 4828 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.214:6443: connect: connection refused" node="crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.222360 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.244735 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: W1210 18:55:20.248558 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-34f2f5fc10ce07fe177ff9a2c0b9ca4b22ef314734e35a7f85e18d7172b9da61 WatchSource:0}: Error finding container 34f2f5fc10ce07fe177ff9a2c0b9ca4b22ef314734e35a7f85e18d7172b9da61: Status 404 returned error can't find the container with id 34f2f5fc10ce07fe177ff9a2c0b9ca4b22ef314734e35a7f85e18d7172b9da61 Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.251253 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.268847 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.273913 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:20 crc kubenswrapper[4828]: W1210 18:55:20.274290 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-44bd78e07cdef32dd225c9cee531442d3709d3bd3ea0563a48c5d84b967cfa15 WatchSource:0}: Error finding container 44bd78e07cdef32dd225c9cee531442d3709d3bd3ea0563a48c5d84b967cfa15: Status 404 returned error can't find the container with id 44bd78e07cdef32dd225c9cee531442d3709d3bd3ea0563a48c5d84b967cfa15 Dec 10 18:55:20 crc kubenswrapper[4828]: W1210 18:55:20.276448 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-3c69c1f880ebc31391b4cd35d6e1f5fe3c0dbebdf0f64835cce1b3128210c910 WatchSource:0}: Error finding container 3c69c1f880ebc31391b4cd35d6e1f5fe3c0dbebdf0f64835cce1b3128210c910: Status 404 returned error can't find the container with id 3c69c1f880ebc31391b4cd35d6e1f5fe3c0dbebdf0f64835cce1b3128210c910 Dec 10 18:55:20 crc kubenswrapper[4828]: W1210 18:55:20.288848 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-beca728b5863a477e28db0a96089c5d333c26f95006fcc1f84729ec5f1325015 WatchSource:0}: Error finding container beca728b5863a477e28db0a96089c5d333c26f95006fcc1f84729ec5f1325015: Status 404 returned error can't find the container with id beca728b5863a477e28db0a96089c5d333c26f95006fcc1f84729ec5f1325015 Dec 10 18:55:20 crc kubenswrapper[4828]: W1210 18:55:20.299870 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-42dfc5d8b7b3ab595a597113b8a9e001921770fad799c77ccce6123190969fd5 WatchSource:0}: Error finding container 42dfc5d8b7b3ab595a597113b8a9e001921770fad799c77ccce6123190969fd5: Status 404 returned error can't find the container with id 42dfc5d8b7b3ab595a597113b8a9e001921770fad799c77ccce6123190969fd5 Dec 10 18:55:20 crc kubenswrapper[4828]: E1210 18:55:20.333156 4828 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="800ms" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.544333 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.545339 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.545376 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.545386 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.545412 4828 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 18:55:20 crc kubenswrapper[4828]: E1210 18:55:20.545906 4828 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.214:6443: connect: connection refused" node="crc" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.726307 4828 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.727274 4828 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 21:13:59.417590116 +0000 UTC Dec 10 18:55:20 crc kubenswrapper[4828]: W1210 18:55:20.740296 4828 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Dec 10 18:55:20 crc kubenswrapper[4828]: E1210 18:55:20.740378 4828 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.791996 4828 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="215812f5a8013b31678b85a9fa5232c2da907e04f587be31f9508ea39182976b" exitCode=0 Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.792075 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"215812f5a8013b31678b85a9fa5232c2da907e04f587be31f9508ea39182976b"} Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.792152 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"34f2f5fc10ce07fe177ff9a2c0b9ca4b22ef314734e35a7f85e18d7172b9da61"} Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.792249 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.793461 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.793504 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.793517 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.794161 4828 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6" exitCode=0 Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.794274 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6"} Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.794323 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"42dfc5d8b7b3ab595a597113b8a9e001921770fad799c77ccce6123190969fd5"} Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.794405 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.795131 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.795152 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.795159 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.796113 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6"} Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.796135 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"beca728b5863a477e28db0a96089c5d333c26f95006fcc1f84729ec5f1325015"} Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.797344 4828 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530" exitCode=0 Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.797385 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530"} Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.797398 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3c69c1f880ebc31391b4cd35d6e1f5fe3c0dbebdf0f64835cce1b3128210c910"} Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.797460 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.798152 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.798197 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.798211 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.800659 4828 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="87cd02d3f27730f570ecb37cc52589437e979208dfa27382c11e92b35314c150" exitCode=0 Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.800681 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"87cd02d3f27730f570ecb37cc52589437e979208dfa27382c11e92b35314c150"} Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.800694 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"44bd78e07cdef32dd225c9cee531442d3709d3bd3ea0563a48c5d84b967cfa15"} Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.800736 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.800818 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.801484 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.801511 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.801521 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.802053 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.802070 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:20 crc kubenswrapper[4828]: I1210 18:55:20.802078 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:20 crc kubenswrapper[4828]: W1210 18:55:20.842723 4828 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Dec 10 18:55:20 crc kubenswrapper[4828]: E1210 18:55:20.842836 4828 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Dec 10 18:55:20 crc kubenswrapper[4828]: W1210 18:55:20.872079 4828 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.214:6443: connect: connection refused Dec 10 18:55:20 crc kubenswrapper[4828]: E1210 18:55:20.872186 4828 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.214:6443: connect: connection refused" logger="UnhandledError" Dec 10 18:55:21 crc kubenswrapper[4828]: E1210 18:55:21.134328 4828 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="1.6s" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.346425 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.347474 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.347515 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.347528 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.347556 4828 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.728786 4828 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 10:46:52.601899448 +0000 UTC Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.806916 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859"} Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.807040 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203"} Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.807082 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9"} Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.807114 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50"} Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.807143 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2"} Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.807209 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.807991 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.808032 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.808042 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.809074 4828 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="05bb84da516cd9d40f541dbb58ea252b3541f3dfa75310f5baa30bbedc69f94c" exitCode=0 Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.809175 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"05bb84da516cd9d40f541dbb58ea252b3541f3dfa75310f5baa30bbedc69f94c"} Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.809395 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.810863 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.810895 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.810904 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.812091 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"6f3f41b839ee9d8b2b5083bd8b5ba463e892011fcfb9052449f9198c6bba5d64"} Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.812296 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.813948 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.814021 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.814050 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.818738 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b044808efab4ceca84bf0256a92f6559ea57b15a0d2a6c83756f489accbefdd8"} Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.818790 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"66fc64d5883f4b746a0636317bd472362758bdbfa1c5ab6a0a327692256e064a"} Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.818829 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b1dcfb31658e3faf516dbebd1e00d1f8dc701e5c6b73831a52648effb712db12"} Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.818907 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.819629 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.819662 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.819678 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.821142 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a"} Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.821175 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e"} Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.821189 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20"} Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.821234 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.821953 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.821991 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.822003 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:21 crc kubenswrapper[4828]: I1210 18:55:21.839571 4828 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.486915 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.729285 4828 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 21:01:48.726316853 +0000 UTC Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.829581 4828 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c256b0c4c4ee71b467155abe20761c346d4663ee4dbd9e93673e25de081c0d2f" exitCode=0 Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.829759 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.829870 4828 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.829896 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c256b0c4c4ee71b467155abe20761c346d4663ee4dbd9e93673e25de081c0d2f"} Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.829959 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.829988 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.830121 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.830179 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.835102 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.835173 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.835195 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.835221 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.835112 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.835290 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.835322 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.835335 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.835338 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.835120 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.835461 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:22 crc kubenswrapper[4828]: I1210 18:55:22.835484 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:23 crc kubenswrapper[4828]: I1210 18:55:23.729982 4828 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 15:52:32.665075168 +0000 UTC Dec 10 18:55:23 crc kubenswrapper[4828]: I1210 18:55:23.807037 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:23 crc kubenswrapper[4828]: I1210 18:55:23.814313 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:23 crc kubenswrapper[4828]: I1210 18:55:23.837645 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7ab69c66776e7962e54cc7e5000bd3d956ea9fddbff4a31d9238a0227fd33015"} Dec 10 18:55:23 crc kubenswrapper[4828]: I1210 18:55:23.837720 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:23 crc kubenswrapper[4828]: I1210 18:55:23.837726 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:23 crc kubenswrapper[4828]: I1210 18:55:23.837728 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d20e43a682f81e00826c1d522fecd6358544b5c2b1efdb361a3715e048303f5d"} Dec 10 18:55:23 crc kubenswrapper[4828]: I1210 18:55:23.837906 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"caabc6495f23767ee72ea3f6b72a7974c56c4c45399f458aced3d1750a9afa29"} Dec 10 18:55:23 crc kubenswrapper[4828]: I1210 18:55:23.838992 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:23 crc kubenswrapper[4828]: I1210 18:55:23.839060 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:23 crc kubenswrapper[4828]: I1210 18:55:23.839070 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:23 crc kubenswrapper[4828]: I1210 18:55:23.839133 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:23 crc kubenswrapper[4828]: I1210 18:55:23.839153 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:23 crc kubenswrapper[4828]: I1210 18:55:23.839087 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.730152 4828 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 10:23:40.905706013 +0000 UTC Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.730222 4828 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 399h28m16.175489829s for next certificate rotation Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.843792 4828 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.843870 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.843980 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1d21fa003fe6665addab3442c1257aa58476f89b89126682102ede4c6af5055c"} Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.844061 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"81ecddf760f632ec655e86d35cb965338c60fc67072fe64cc45b40c68c0653e6"} Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.844000 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.844894 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.844936 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.844953 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.845241 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.845294 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.845311 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.929323 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.929487 4828 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.929544 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.930922 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.930968 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:24 crc kubenswrapper[4828]: I1210 18:55:24.930984 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:25 crc kubenswrapper[4828]: I1210 18:55:25.062276 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:25 crc kubenswrapper[4828]: I1210 18:55:25.416208 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:25 crc kubenswrapper[4828]: I1210 18:55:25.738363 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:25 crc kubenswrapper[4828]: I1210 18:55:25.845675 4828 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 18:55:25 crc kubenswrapper[4828]: I1210 18:55:25.845742 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:25 crc kubenswrapper[4828]: I1210 18:55:25.845887 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:25 crc kubenswrapper[4828]: I1210 18:55:25.845923 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:25 crc kubenswrapper[4828]: I1210 18:55:25.846565 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:25 crc kubenswrapper[4828]: I1210 18:55:25.846597 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:25 crc kubenswrapper[4828]: I1210 18:55:25.846606 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:25 crc kubenswrapper[4828]: I1210 18:55:25.846913 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:25 crc kubenswrapper[4828]: I1210 18:55:25.846948 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:25 crc kubenswrapper[4828]: I1210 18:55:25.846959 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:25 crc kubenswrapper[4828]: I1210 18:55:25.847181 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:25 crc kubenswrapper[4828]: I1210 18:55:25.847212 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:25 crc kubenswrapper[4828]: I1210 18:55:25.847225 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:26 crc kubenswrapper[4828]: I1210 18:55:26.076205 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:26 crc kubenswrapper[4828]: I1210 18:55:26.848280 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:26 crc kubenswrapper[4828]: I1210 18:55:26.848291 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:26 crc kubenswrapper[4828]: I1210 18:55:26.849577 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:26 crc kubenswrapper[4828]: I1210 18:55:26.849647 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:26 crc kubenswrapper[4828]: I1210 18:55:26.849670 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:26 crc kubenswrapper[4828]: I1210 18:55:26.849937 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:26 crc kubenswrapper[4828]: I1210 18:55:26.849964 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:26 crc kubenswrapper[4828]: I1210 18:55:26.849976 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:26 crc kubenswrapper[4828]: I1210 18:55:26.923753 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 10 18:55:26 crc kubenswrapper[4828]: I1210 18:55:26.923963 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:26 crc kubenswrapper[4828]: I1210 18:55:26.925221 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:26 crc kubenswrapper[4828]: I1210 18:55:26.925276 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:26 crc kubenswrapper[4828]: I1210 18:55:26.925298 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:27 crc kubenswrapper[4828]: I1210 18:55:27.372825 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 10 18:55:27 crc kubenswrapper[4828]: I1210 18:55:27.851348 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:27 crc kubenswrapper[4828]: I1210 18:55:27.852789 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:27 crc kubenswrapper[4828]: I1210 18:55:27.852957 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:27 crc kubenswrapper[4828]: I1210 18:55:27.852985 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:28 crc kubenswrapper[4828]: I1210 18:55:28.416415 4828 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 10 18:55:28 crc kubenswrapper[4828]: I1210 18:55:28.416524 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 18:55:29 crc kubenswrapper[4828]: E1210 18:55:29.847176 4828 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 10 18:55:31 crc kubenswrapper[4828]: W1210 18:55:31.322997 4828 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 10 18:55:31 crc kubenswrapper[4828]: I1210 18:55:31.323109 4828 trace.go:236] Trace[490371809]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Dec-2025 18:55:21.320) (total time: 10002ms): Dec 10 18:55:31 crc kubenswrapper[4828]: Trace[490371809]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (18:55:31.322) Dec 10 18:55:31 crc kubenswrapper[4828]: Trace[490371809]: [10.002224517s] [10.002224517s] END Dec 10 18:55:31 crc kubenswrapper[4828]: E1210 18:55:31.323133 4828 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 10 18:55:31 crc kubenswrapper[4828]: E1210 18:55:31.349222 4828 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 10 18:55:31 crc kubenswrapper[4828]: I1210 18:55:31.726424 4828 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 10 18:55:31 crc kubenswrapper[4828]: E1210 18:55:31.841910 4828 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 10 18:55:32 crc kubenswrapper[4828]: E1210 18:55:32.734897 4828 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 10 18:55:32 crc kubenswrapper[4828]: W1210 18:55:32.760097 4828 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 10 18:55:32 crc kubenswrapper[4828]: I1210 18:55:32.760352 4828 trace.go:236] Trace[1985613277]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Dec-2025 18:55:22.758) (total time: 10002ms): Dec 10 18:55:32 crc kubenswrapper[4828]: Trace[1985613277]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (18:55:32.760) Dec 10 18:55:32 crc kubenswrapper[4828]: Trace[1985613277]: [10.002035502s] [10.002035502s] END Dec 10 18:55:32 crc kubenswrapper[4828]: E1210 18:55:32.760488 4828 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 10 18:55:32 crc kubenswrapper[4828]: I1210 18:55:32.902708 4828 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 10 18:55:32 crc kubenswrapper[4828]: I1210 18:55:32.903133 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 10 18:55:32 crc kubenswrapper[4828]: I1210 18:55:32.909110 4828 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 10 18:55:32 crc kubenswrapper[4828]: I1210 18:55:32.909489 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 10 18:55:32 crc kubenswrapper[4828]: I1210 18:55:32.949331 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:32 crc kubenswrapper[4828]: I1210 18:55:32.951141 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:32 crc kubenswrapper[4828]: I1210 18:55:32.951182 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:32 crc kubenswrapper[4828]: I1210 18:55:32.951194 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:32 crc kubenswrapper[4828]: I1210 18:55:32.951223 4828 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 18:55:34 crc kubenswrapper[4828]: I1210 18:55:34.376874 4828 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 10 18:55:34 crc kubenswrapper[4828]: I1210 18:55:34.936650 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:34 crc kubenswrapper[4828]: I1210 18:55:34.936770 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:34 crc kubenswrapper[4828]: I1210 18:55:34.937543 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:34 crc kubenswrapper[4828]: I1210 18:55:34.937563 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:34 crc kubenswrapper[4828]: I1210 18:55:34.937573 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:34 crc kubenswrapper[4828]: I1210 18:55:34.942770 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:35 crc kubenswrapper[4828]: I1210 18:55:35.869885 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:35 crc kubenswrapper[4828]: I1210 18:55:35.871537 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:35 crc kubenswrapper[4828]: I1210 18:55:35.871611 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:35 crc kubenswrapper[4828]: I1210 18:55:35.871625 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:35 crc kubenswrapper[4828]: I1210 18:55:35.918779 4828 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 10 18:55:35 crc kubenswrapper[4828]: I1210 18:55:35.935620 4828 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 10 18:55:36 crc kubenswrapper[4828]: I1210 18:55:36.080424 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:36 crc kubenswrapper[4828]: I1210 18:55:36.080644 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:36 crc kubenswrapper[4828]: I1210 18:55:36.082204 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:36 crc kubenswrapper[4828]: I1210 18:55:36.082276 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:36 crc kubenswrapper[4828]: I1210 18:55:36.082296 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:36 crc kubenswrapper[4828]: I1210 18:55:36.498192 4828 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 10 18:55:36 crc kubenswrapper[4828]: I1210 18:55:36.948442 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 10 18:55:36 crc kubenswrapper[4828]: I1210 18:55:36.948941 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:36 crc kubenswrapper[4828]: I1210 18:55:36.949892 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:36 crc kubenswrapper[4828]: I1210 18:55:36.949920 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:36 crc kubenswrapper[4828]: I1210 18:55:36.949929 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:36 crc kubenswrapper[4828]: I1210 18:55:36.967026 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 10 18:55:37 crc kubenswrapper[4828]: I1210 18:55:37.875636 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:37 crc kubenswrapper[4828]: I1210 18:55:37.877092 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:37 crc kubenswrapper[4828]: I1210 18:55:37.877153 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:37 crc kubenswrapper[4828]: I1210 18:55:37.877169 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:37 crc kubenswrapper[4828]: I1210 18:55:37.905324 4828 trace.go:236] Trace[2078641928]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Dec-2025 18:55:23.633) (total time: 14271ms): Dec 10 18:55:37 crc kubenswrapper[4828]: Trace[2078641928]: ---"Objects listed" error: 14271ms (18:55:37.905) Dec 10 18:55:37 crc kubenswrapper[4828]: Trace[2078641928]: [14.271431682s] [14.271431682s] END Dec 10 18:55:37 crc kubenswrapper[4828]: I1210 18:55:37.905386 4828 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 10 18:55:37 crc kubenswrapper[4828]: I1210 18:55:37.906506 4828 trace.go:236] Trace[578895664]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (10-Dec-2025 18:55:23.796) (total time: 14109ms): Dec 10 18:55:37 crc kubenswrapper[4828]: Trace[578895664]: ---"Objects listed" error: 14109ms (18:55:37.906) Dec 10 18:55:37 crc kubenswrapper[4828]: Trace[578895664]: [14.109890802s] [14.109890802s] END Dec 10 18:55:37 crc kubenswrapper[4828]: I1210 18:55:37.906564 4828 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 10 18:55:37 crc kubenswrapper[4828]: I1210 18:55:37.909172 4828 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 10 18:55:37 crc kubenswrapper[4828]: I1210 18:55:37.931736 4828 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37910->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 10 18:55:37 crc kubenswrapper[4828]: I1210 18:55:37.932001 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37910->192.168.126.11:17697: read: connection reset by peer" Dec 10 18:55:37 crc kubenswrapper[4828]: I1210 18:55:37.932583 4828 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 10 18:55:37 crc kubenswrapper[4828]: I1210 18:55:37.932636 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 10 18:55:37 crc kubenswrapper[4828]: I1210 18:55:37.933138 4828 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 10 18:55:37 crc kubenswrapper[4828]: I1210 18:55:37.933233 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 10 18:55:37 crc kubenswrapper[4828]: E1210 18:55:37.966752 4828 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 10 18:55:37 crc kubenswrapper[4828]: I1210 18:55:37.989773 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:37 crc kubenswrapper[4828]: I1210 18:55:37.997289 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.362588 4828 csr.go:261] certificate signing request csr-wxxrt is approved, waiting to be issued Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.369117 4828 csr.go:257] certificate signing request csr-wxxrt is issued Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.729051 4828 apiserver.go:52] "Watching apiserver" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.730572 4828 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.730611 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.733353 4828 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.733645 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.734013 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.734110 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.734155 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:38 crc kubenswrapper[4828]: E1210 18:55:38.734279 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.734460 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:38 crc kubenswrapper[4828]: E1210 18:55:38.734501 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.734517 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:38 crc kubenswrapper[4828]: E1210 18:55:38.734645 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.735004 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.738429 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.738442 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.738786 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.738946 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.744548 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.744593 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.744599 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.745312 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.745593 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.805866 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-hnf4q"] Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.806133 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hnf4q" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.813756 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.814065 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.814111 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.829754 4828 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.849531 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.866710 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.878656 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.880482 4828 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859" exitCode=255 Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.880561 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859"} Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.886206 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:38 crc kubenswrapper[4828]: E1210 18:55:38.888349 4828 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.896840 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.904258 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.912828 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.914027 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.914149 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.914246 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.914323 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.914373 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.914393 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.914525 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.914461 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.914563 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.914690 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.914761 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.914854 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.914791 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.914922 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915037 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915069 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915093 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915116 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915139 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915161 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915183 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915201 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915224 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915245 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915265 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915291 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915312 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915337 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915357 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915375 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915394 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915413 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915431 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915450 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915475 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915511 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915531 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915554 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915574 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915606 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915627 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915650 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915675 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915694 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915713 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915732 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915751 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915773 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915791 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915826 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915849 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915866 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915884 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915901 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915919 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915936 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915955 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915976 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915994 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916014 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916031 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916048 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916067 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916085 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916104 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916123 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916188 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916208 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916228 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916243 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916261 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916278 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916295 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916313 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916328 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916345 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916367 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916385 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916403 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916420 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916438 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916459 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916476 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916492 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916509 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916526 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916544 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916561 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916576 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916594 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916612 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916630 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916648 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916666 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916683 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916699 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.917812 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.917848 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.917869 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.917912 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.917929 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.917948 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.917966 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.917983 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918001 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918019 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918037 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918052 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918068 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918084 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918100 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918118 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918741 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918793 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918827 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918846 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918866 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918887 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918906 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918926 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918948 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918970 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918988 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919005 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919023 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919044 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919062 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919079 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919097 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919118 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919136 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919158 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919174 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919194 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919211 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919229 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919254 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919311 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919331 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915252 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919350 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915492 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919370 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915500 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915580 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915709 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915717 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915924 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915941 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915956 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.915970 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916014 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916085 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916118 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916144 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916152 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916233 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916412 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916452 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916600 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916818 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916845 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916849 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.916891 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.917014 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.917032 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.917058 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.917122 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.917289 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.917463 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.917524 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.917616 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.917762 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918383 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918408 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918635 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918782 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.918866 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919015 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919081 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919152 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919337 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919477 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919504 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919537 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919555 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919730 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919847 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919858 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919878 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.920708 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.920718 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.920750 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.921014 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.921037 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.921497 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.921548 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.921734 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.921758 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.921893 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.922030 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.922101 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.922159 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.922247 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.922397 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.923360 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.923466 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.923931 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.923966 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.924045 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.924112 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.924119 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.924404 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.924444 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.924555 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.924690 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.919393 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.925020 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.925066 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.925112 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.925900 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.925926 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.925949 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926013 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926032 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926054 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926076 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926102 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926120 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926140 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926157 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926175 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926194 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926211 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926227 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926245 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926262 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926280 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926298 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926315 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926331 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926349 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926366 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926383 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926522 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926546 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926562 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926580 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926595 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926611 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926629 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926644 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926661 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926679 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926702 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926719 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926735 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926756 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926774 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926792 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926832 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926849 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926865 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926898 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926915 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926932 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926949 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926966 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926987 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927011 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927029 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927046 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927062 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927077 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927093 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927136 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927160 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927179 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927202 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927226 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927248 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927266 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/74b63880-331e-46a7-895b-434b52ff35b2-hosts-file\") pod \"node-resolver-hnf4q\" (UID: \"74b63880-331e-46a7-895b-434b52ff35b2\") " pod="openshift-dns/node-resolver-hnf4q" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927285 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtbf8\" (UniqueName: \"kubernetes.io/projected/74b63880-331e-46a7-895b-434b52ff35b2-kube-api-access-mtbf8\") pod \"node-resolver-hnf4q\" (UID: \"74b63880-331e-46a7-895b-434b52ff35b2\") " pod="openshift-dns/node-resolver-hnf4q" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927302 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927322 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927341 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927361 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927380 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927397 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927418 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927435 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927507 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927518 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927530 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927539 4828 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927548 4828 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927558 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927567 4828 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927576 4828 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927587 4828 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927598 4828 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927608 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927617 4828 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927626 4828 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927634 4828 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927643 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927653 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927662 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927671 4828 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927681 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927690 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927699 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927709 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927719 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927730 4828 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927739 4828 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927748 4828 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927756 4828 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927764 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927774 4828 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927783 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927792 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927821 4828 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927835 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927843 4828 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927853 4828 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927862 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927871 4828 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927880 4828 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927889 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927901 4828 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927910 4828 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927919 4828 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927929 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927938 4828 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927947 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927956 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927965 4828 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927973 4828 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927982 4828 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927991 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928000 4828 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928010 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928018 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928028 4828 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928038 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928046 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928056 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928064 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928073 4828 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928082 4828 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928092 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928101 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928109 4828 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928117 4828 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928127 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928136 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928145 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928154 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928163 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928172 4828 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928181 4828 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928192 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928200 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928209 4828 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928218 4828 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928227 4828 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928236 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928246 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928255 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.925338 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.925373 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.925389 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.925554 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.925690 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.925872 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.925939 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926007 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926149 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926263 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926477 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926748 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.926851 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927605 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927635 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928365 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927667 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.927907 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928476 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928135 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928607 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928626 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928647 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928917 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.928927 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.929132 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.929336 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.929704 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.930072 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.930561 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.930862 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.930979 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.931006 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.931064 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.931372 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.931497 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.931512 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.931566 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.930506 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.931615 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.931550 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: E1210 18:55:38.931909 4828 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.931958 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: E1210 18:55:38.932137 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:39.432009067 +0000 UTC m=+19.942620152 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.932269 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.932284 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.932617 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.932631 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.932873 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.932887 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.933009 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.933031 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.933183 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.933260 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.933325 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.933326 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.933382 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.933569 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.933617 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.933781 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.933946 4828 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.934013 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.934103 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.934180 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:38 crc kubenswrapper[4828]: E1210 18:55:38.934281 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:55:39.434259729 +0000 UTC m=+19.944870824 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.934337 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.934451 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.934505 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.934695 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.934698 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.934956 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.935062 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.935145 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.935390 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.935479 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.935599 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: E1210 18:55:38.935820 4828 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.935836 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.935866 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: E1210 18:55:38.935887 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:39.43586978 +0000 UTC m=+19.946480785 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.935895 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.935910 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.935881 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.936156 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.936179 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.936303 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.936315 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.936324 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.936366 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.937332 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.937909 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.938269 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.938456 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.938516 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.938774 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.939213 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.939877 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.940251 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.946245 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.947154 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.947232 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.947261 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.947351 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.947503 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.949455 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.950015 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.950081 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: E1210 18:55:38.950505 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:38 crc kubenswrapper[4828]: E1210 18:55:38.950576 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:38 crc kubenswrapper[4828]: E1210 18:55:38.950593 4828 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.950608 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: E1210 18:55:38.950675 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:39.450650703 +0000 UTC m=+19.961261948 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.950827 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.951025 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.952165 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.953047 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.953107 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:38 crc kubenswrapper[4828]: E1210 18:55:38.953322 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:38 crc kubenswrapper[4828]: E1210 18:55:38.953343 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:38 crc kubenswrapper[4828]: E1210 18:55:38.953353 4828 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:38 crc kubenswrapper[4828]: E1210 18:55:38.953383 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:39.45337258 +0000 UTC m=+19.963983585 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.953909 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.954539 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.956026 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.957271 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.957357 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.957447 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.962317 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.962599 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.963971 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.964356 4828 scope.go:117] "RemoveContainer" containerID="bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.964921 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.965421 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.965453 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.965604 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.966539 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.966671 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.970976 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.984856 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.986705 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.989268 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:38 crc kubenswrapper[4828]: I1210 18:55:38.990023 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.000026 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.007463 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.012002 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.022077 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.029639 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.029709 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.029742 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/74b63880-331e-46a7-895b-434b52ff35b2-hosts-file\") pod \"node-resolver-hnf4q\" (UID: \"74b63880-331e-46a7-895b-434b52ff35b2\") " pod="openshift-dns/node-resolver-hnf4q" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.029767 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtbf8\" (UniqueName: \"kubernetes.io/projected/74b63880-331e-46a7-895b-434b52ff35b2-kube-api-access-mtbf8\") pod \"node-resolver-hnf4q\" (UID: \"74b63880-331e-46a7-895b-434b52ff35b2\") " pod="openshift-dns/node-resolver-hnf4q" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.029839 4828 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.029857 4828 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.029871 4828 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.029883 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.029895 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.029908 4828 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.029920 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.029933 4828 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.029944 4828 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.029956 4828 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.029968 4828 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.029979 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.029992 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030003 4828 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030030 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030040 4828 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030048 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030018 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030050 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030153 4828 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030155 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/74b63880-331e-46a7-895b-434b52ff35b2-hosts-file\") pod \"node-resolver-hnf4q\" (UID: \"74b63880-331e-46a7-895b-434b52ff35b2\") " pod="openshift-dns/node-resolver-hnf4q" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030164 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030215 4828 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030227 4828 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030237 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030246 4828 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030255 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030264 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030273 4828 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030282 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030292 4828 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030300 4828 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030309 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030318 4828 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030326 4828 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030335 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030343 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030368 4828 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030376 4828 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030385 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030394 4828 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030402 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030411 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030419 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030428 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030437 4828 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030446 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030454 4828 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030464 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030473 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030485 4828 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030495 4828 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030506 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030516 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030525 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030534 4828 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030543 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030551 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030560 4828 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030569 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030577 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030586 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030594 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030602 4828 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030610 4828 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030618 4828 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030627 4828 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030636 4828 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030644 4828 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030652 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030661 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030670 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030678 4828 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030687 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030696 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030704 4828 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030713 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030723 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030731 4828 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030739 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030748 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030757 4828 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030766 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030775 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030784 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030832 4828 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030842 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030852 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.030860 4828 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031077 4828 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031088 4828 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031098 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031107 4828 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031117 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031126 4828 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031134 4828 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031142 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031151 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031165 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031174 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031184 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031192 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031200 4828 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031209 4828 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031219 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031255 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031265 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031274 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031296 4828 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031305 4828 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031313 4828 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031321 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031331 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031339 4828 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031348 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031356 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031365 4828 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.031424 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.040817 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.045958 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtbf8\" (UniqueName: \"kubernetes.io/projected/74b63880-331e-46a7-895b-434b52ff35b2-kube-api-access-mtbf8\") pod \"node-resolver-hnf4q\" (UID: \"74b63880-331e-46a7-895b-434b52ff35b2\") " pod="openshift-dns/node-resolver-hnf4q" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.047362 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.048943 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.059015 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.059444 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.065657 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.116657 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hnf4q" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.255965 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-rzndz"] Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.257029 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.261925 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-677ql"] Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.262439 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.263536 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.265857 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.266110 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.266952 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.267488 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-xdxgd"] Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.267846 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8f42x"] Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.268726 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.269171 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.283330 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.288547 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.288874 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.288990 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.289050 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.289092 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.289229 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.289346 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.289449 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.289742 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.289879 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.289989 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.290168 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.290437 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.290584 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.323218 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335161 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-var-lib-cni-multus\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335199 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-hostroot\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335221 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hdv4\" (UniqueName: \"kubernetes.io/projected/1f37f60f-fe12-4530-8ff1-ea3794c65597-kube-api-access-4hdv4\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335237 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-run-netns\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335259 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/40784d67-bada-4229-8764-efba7c96edb5-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335291 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-system-cni-dir\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335306 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-log-socket\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335321 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-run-ovn-kubernetes\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335339 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-env-overrides\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335353 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c6e55203-65ae-4767-9a96-f309bc41ddf3-multus-daemon-config\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335368 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-slash\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335380 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-run-netns\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335394 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-cni-bin\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335408 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-etc-openvswitch\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335451 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-systemd\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335468 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-multus-socket-dir-parent\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335484 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovn-node-metrics-cert\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335498 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c6e55203-65ae-4767-9a96-f309bc41ddf3-cni-binary-copy\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335514 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/40784d67-bada-4229-8764-efba7c96edb5-cni-binary-copy\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.335530 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c0bc5145-c8c4-46a3-adf1-5d64ac422dc4-mcd-auth-proxy-config\") pod \"machine-config-daemon-677ql\" (UID: \"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\") " pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.336876 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-node-log\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.337094 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-os-release\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.337119 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-multus-conf-dir\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.341912 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/40784d67-bada-4229-8764-efba7c96edb5-system-cni-dir\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.341998 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c0bc5145-c8c4-46a3-adf1-5d64ac422dc4-rootfs\") pod \"machine-config-daemon-677ql\" (UID: \"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\") " pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342052 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-systemd-units\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342115 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/40784d67-bada-4229-8764-efba7c96edb5-cnibin\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342136 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-var-lib-openvswitch\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342162 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-openvswitch\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342195 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342249 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-var-lib-cni-bin\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342351 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-ovn\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342408 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovnkube-script-lib\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342435 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-var-lib-kubelet\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342482 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-run-multus-certs\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342512 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c0bc5145-c8c4-46a3-adf1-5d64ac422dc4-proxy-tls\") pod \"machine-config-daemon-677ql\" (UID: \"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\") " pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342566 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-cnibin\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342600 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-run-k8s-cni-cncf-io\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342634 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/40784d67-bada-4229-8764-efba7c96edb5-os-release\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342676 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-etc-kubernetes\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342705 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srfr2\" (UniqueName: \"kubernetes.io/projected/c6e55203-65ae-4767-9a96-f309bc41ddf3-kube-api-access-srfr2\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342735 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsnvv\" (UniqueName: \"kubernetes.io/projected/40784d67-bada-4229-8764-efba7c96edb5-kube-api-access-wsnvv\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342773 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-kubelet\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342838 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-multus-cni-dir\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342871 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/40784d67-bada-4229-8764-efba7c96edb5-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342902 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp2lf\" (UniqueName: \"kubernetes.io/projected/c0bc5145-c8c4-46a3-adf1-5d64ac422dc4-kube-api-access-qp2lf\") pod \"machine-config-daemon-677ql\" (UID: \"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\") " pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.342932 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-cni-netd\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.343019 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovnkube-config\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.370765 4828 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-10 18:50:38 +0000 UTC, rotation deadline is 2026-10-01 09:37:13.352088712 +0000 UTC Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.370829 4828 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7070h41m33.981261872s for next certificate rotation Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.370870 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.384049 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.408985 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.417350 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.429671 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.439687 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.443978 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444058 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-ovn\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444083 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovnkube-script-lib\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: E1210 18:55:39.444109 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:55:40.444085402 +0000 UTC m=+20.954696437 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444164 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-var-lib-kubelet\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444180 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-ovn\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444195 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-run-multus-certs\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444243 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c0bc5145-c8c4-46a3-adf1-5d64ac422dc4-proxy-tls\") pod \"machine-config-daemon-677ql\" (UID: \"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\") " pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444254 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-var-lib-kubelet\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444264 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-cnibin\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444290 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-run-k8s-cni-cncf-io\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444333 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-run-k8s-cni-cncf-io\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444370 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/40784d67-bada-4229-8764-efba7c96edb5-os-release\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444452 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-etc-kubernetes\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444477 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srfr2\" (UniqueName: \"kubernetes.io/projected/c6e55203-65ae-4767-9a96-f309bc41ddf3-kube-api-access-srfr2\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444506 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsnvv\" (UniqueName: \"kubernetes.io/projected/40784d67-bada-4229-8764-efba7c96edb5-kube-api-access-wsnvv\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444530 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-kubelet\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444540 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-etc-kubernetes\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444391 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-run-multus-certs\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444569 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/40784d67-bada-4229-8764-efba7c96edb5-os-release\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444550 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/40784d67-bada-4229-8764-efba7c96edb5-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444394 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-cnibin\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444606 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-kubelet\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444655 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp2lf\" (UniqueName: \"kubernetes.io/projected/c0bc5145-c8c4-46a3-adf1-5d64ac422dc4-kube-api-access-qp2lf\") pod \"machine-config-daemon-677ql\" (UID: \"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\") " pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444667 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovnkube-script-lib\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444685 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-cni-netd\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444711 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-multus-cni-dir\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444738 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444760 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovnkube-config\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444785 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hdv4\" (UniqueName: \"kubernetes.io/projected/1f37f60f-fe12-4530-8ff1-ea3794c65597-kube-api-access-4hdv4\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444763 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-cni-netd\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444834 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/40784d67-bada-4229-8764-efba7c96edb5-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: E1210 18:55:39.444824 4828 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444933 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-multus-cni-dir\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: E1210 18:55:39.444945 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:40.444919489 +0000 UTC m=+20.955530584 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444843 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-run-netns\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444824 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-run-netns\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.444983 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-var-lib-cni-multus\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445005 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-hostroot\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445029 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/40784d67-bada-4229-8764-efba7c96edb5-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445049 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-system-cni-dir\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445068 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-log-socket\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445082 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-hostroot\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445119 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-run-ovn-kubernetes\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445087 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-run-ovn-kubernetes\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445166 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-var-lib-cni-multus\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445183 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-env-overrides\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445202 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-slash\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445212 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-system-cni-dir\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445219 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-run-netns\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445235 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-cni-bin\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445246 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-log-socket\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445255 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c6e55203-65ae-4767-9a96-f309bc41ddf3-multus-daemon-config\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445270 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovnkube-config\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445276 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-slash\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445291 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-etc-openvswitch\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445271 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-etc-openvswitch\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445314 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-run-netns\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445331 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445341 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-cni-bin\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445358 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-systemd\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445374 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-multus-socket-dir-parent\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445389 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/40784d67-bada-4229-8764-efba7c96edb5-cni-binary-copy\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445411 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c0bc5145-c8c4-46a3-adf1-5d64ac422dc4-mcd-auth-proxy-config\") pod \"machine-config-daemon-677ql\" (UID: \"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\") " pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445424 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-node-log\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445438 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovn-node-metrics-cert\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445465 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c6e55203-65ae-4767-9a96-f309bc41ddf3-cni-binary-copy\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445501 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/40784d67-bada-4229-8764-efba7c96edb5-system-cni-dir\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445526 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c0bc5145-c8c4-46a3-adf1-5d64ac422dc4-rootfs\") pod \"machine-config-daemon-677ql\" (UID: \"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\") " pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445541 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-os-release\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445557 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-multus-conf-dir\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445575 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-systemd-units\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445601 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/40784d67-bada-4229-8764-efba7c96edb5-cnibin\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445648 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-var-lib-openvswitch\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445663 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-openvswitch\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445680 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445698 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-var-lib-cni-bin\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445816 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-host-var-lib-cni-bin\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445824 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-env-overrides\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: E1210 18:55:39.445863 4828 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445869 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/40784d67-bada-4229-8764-efba7c96edb5-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445881 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/40784d67-bada-4229-8764-efba7c96edb5-system-cni-dir\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: E1210 18:55:39.445897 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:40.44588858 +0000 UTC m=+20.956499575 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445910 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c6e55203-65ae-4767-9a96-f309bc41ddf3-multus-daemon-config\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445913 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-systemd\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445930 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c0bc5145-c8c4-46a3-adf1-5d64ac422dc4-rootfs\") pod \"machine-config-daemon-677ql\" (UID: \"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\") " pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445941 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-node-log\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445972 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-os-release\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.445987 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-multus-socket-dir-parent\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.446001 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c6e55203-65ae-4767-9a96-f309bc41ddf3-multus-conf-dir\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.446020 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-var-lib-openvswitch\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.446033 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-openvswitch\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.446047 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-systemd-units\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.446058 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.446072 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/40784d67-bada-4229-8764-efba7c96edb5-cnibin\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.446554 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/40784d67-bada-4229-8764-efba7c96edb5-cni-binary-copy\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.446573 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c0bc5145-c8c4-46a3-adf1-5d64ac422dc4-mcd-auth-proxy-config\") pod \"machine-config-daemon-677ql\" (UID: \"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\") " pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.446722 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c6e55203-65ae-4767-9a96-f309bc41ddf3-cni-binary-copy\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.448908 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c0bc5145-c8c4-46a3-adf1-5d64ac422dc4-proxy-tls\") pod \"machine-config-daemon-677ql\" (UID: \"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\") " pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.448930 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovn-node-metrics-cert\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.460156 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.463869 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp2lf\" (UniqueName: \"kubernetes.io/projected/c0bc5145-c8c4-46a3-adf1-5d64ac422dc4-kube-api-access-qp2lf\") pod \"machine-config-daemon-677ql\" (UID: \"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\") " pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.464573 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hdv4\" (UniqueName: \"kubernetes.io/projected/1f37f60f-fe12-4530-8ff1-ea3794c65597-kube-api-access-4hdv4\") pod \"ovnkube-node-8f42x\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.464620 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsnvv\" (UniqueName: \"kubernetes.io/projected/40784d67-bada-4229-8764-efba7c96edb5-kube-api-access-wsnvv\") pod \"multus-additional-cni-plugins-rzndz\" (UID: \"40784d67-bada-4229-8764-efba7c96edb5\") " pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.467044 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srfr2\" (UniqueName: \"kubernetes.io/projected/c6e55203-65ae-4767-9a96-f309bc41ddf3-kube-api-access-srfr2\") pod \"multus-xdxgd\" (UID: \"c6e55203-65ae-4767-9a96-f309bc41ddf3\") " pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.473553 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.492636 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.505919 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.515980 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.531632 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.546262 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.546308 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:39 crc kubenswrapper[4828]: E1210 18:55:39.546400 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:39 crc kubenswrapper[4828]: E1210 18:55:39.546415 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:39 crc kubenswrapper[4828]: E1210 18:55:39.546425 4828 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:39 crc kubenswrapper[4828]: E1210 18:55:39.546400 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:39 crc kubenswrapper[4828]: E1210 18:55:39.546463 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:40.546450287 +0000 UTC m=+21.057061292 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:39 crc kubenswrapper[4828]: E1210 18:55:39.546470 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:39 crc kubenswrapper[4828]: E1210 18:55:39.546505 4828 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:39 crc kubenswrapper[4828]: E1210 18:55:39.546554 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:40.54654539 +0000 UTC m=+21.057156395 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.549367 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.560585 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.570044 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.579585 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.589124 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.596620 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.606869 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rzndz" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.610329 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.624351 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.625238 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.637933 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0bc5145_c8c4_46a3_adf1_5d64ac422dc4.slice/crio-93cc7f41778b13e5e73f961b7b8b5c99105d8c268c01c421e3321f52e5a98a12 WatchSource:0}: Error finding container 93cc7f41778b13e5e73f961b7b8b5c99105d8c268c01c421e3321f52e5a98a12: Status 404 returned error can't find the container with id 93cc7f41778b13e5e73f961b7b8b5c99105d8c268c01c421e3321f52e5a98a12 Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.638098 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.638074 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.648177 4828 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.648357 4828 reflector.go:484] object-"openshift-network-node-identity"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.648403 4828 reflector.go:484] object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.648725 4828 reflector.go:484] object-"openshift-network-node-identity"/"ovnkube-identity-cm": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"ovnkube-identity-cm": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.648772 4828 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.648822 4828 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.648950 4828 reflector.go:484] object-"openshift-network-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: E1210 18:55:39.648877 4828 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/events\": read tcp 38.102.83.214:53766->38.102.83.214:6443: use of closed network connection" event="&Event{ObjectMeta:{machine-config-daemon-677ql.187fef86706dc132 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:machine-config-daemon-677ql,UID:c0bc5145-c8c4-46a3-adf1-5d64ac422dc4,APIVersion:v1,ResourceVersion:26637,FieldPath:spec.containers{machine-config-daemon},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 18:55:39.643420978 +0000 UTC m=+20.154031983,LastTimestamp:2025-12-10 18:55:39.643420978 +0000 UTC m=+20.154031983,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.648981 4828 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovnkube-script-lib": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovnkube-script-lib": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649000 4828 reflector.go:484] object-"openshift-network-operator"/"metrics-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-network-operator"/"metrics-tls": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649011 4828 reflector.go:484] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649020 4828 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649036 4828 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": watch of *v1.Secret ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649050 4828 reflector.go:484] object-"openshift-machine-config-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649060 4828 reflector.go:484] object-"openshift-multus"/"multus-daemon-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"multus-daemon-config": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649067 4828 reflector.go:484] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649082 4828 reflector.go:484] object-"openshift-multus"/"default-dockercfg-2q5b6": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"default-dockercfg-2q5b6": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649096 4828 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649114 4828 reflector.go:484] object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649132 4828 reflector.go:484] object-"openshift-machine-config-operator"/"proxy-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"proxy-tls": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649151 4828 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovnkube-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovnkube-config": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649194 4828 reflector.go:484] object-"openshift-network-node-identity"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649249 4828 reflector.go:484] object-"openshift-machine-config-operator"/"kube-rbac-proxy": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-rbac-proxy": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649276 4828 reflector.go:484] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649307 4828 reflector.go:484] object-"openshift-network-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649326 4828 reflector.go:484] object-"openshift-network-node-identity"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649342 4828 reflector.go:484] object-"openshift-network-node-identity"/"network-node-identity-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-network-node-identity"/"network-node-identity-cert": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649358 4828 reflector.go:484] object-"openshift-multus"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.649372 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Patch \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-multus/pods/multus-xdxgd/status\": read tcp 38.102.83.214:53766->38.102.83.214:6443: use of closed network connection" Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649445 4828 reflector.go:484] object-"openshift-multus"/"default-cni-sysctl-allowlist": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"default-cni-sysctl-allowlist": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649473 4828 reflector.go:484] object-"openshift-ovn-kubernetes"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649038 4828 reflector.go:484] object-"openshift-multus"/"cni-copy-resources": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"cni-copy-resources": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649084 4828 reflector.go:484] object-"openshift-multus"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.649511 4828 reflector.go:484] object-"openshift-network-operator"/"iptables-alerter-script": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"iptables-alerter-script": Unexpected watch close - watch lasted less than a second and no items received Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.654995 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f37f60f_fe12_4530_8ff1_ea3794c65597.slice/crio-d9089e51b51e3ffe9659d4b0684eb789eb5eeaa74d610fc958e741c6c6637568 WatchSource:0}: Error finding container d9089e51b51e3ffe9659d4b0684eb789eb5eeaa74d610fc958e741c6c6637568: Status 404 returned error can't find the container with id d9089e51b51e3ffe9659d4b0684eb789eb5eeaa74d610fc958e741c6c6637568 Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.658063 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xdxgd" Dec 10 18:55:39 crc kubenswrapper[4828]: W1210 18:55:39.688212 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6e55203_65ae_4767_9a96_f309bc41ddf3.slice/crio-abdf024b4a7dfa923589012b4b6cd2999144df633c29e7445ead216194105b7f WatchSource:0}: Error finding container abdf024b4a7dfa923589012b4b6cd2999144df633c29e7445ead216194105b7f: Status 404 returned error can't find the container with id abdf024b4a7dfa923589012b4b6cd2999144df633c29e7445ead216194105b7f Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.792506 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.793046 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.794453 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.795119 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.796542 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.798155 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.798861 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.800306 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.800787 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.801307 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.802277 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.802758 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.803870 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.805381 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.806214 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.807325 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.807876 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.808883 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.809283 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.809918 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.813481 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.814118 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.815276 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.815769 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.816029 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.816986 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.817418 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.818195 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.819405 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.820167 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.821312 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.821843 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.822702 4828 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.822835 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.824672 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.825991 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.826474 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.828780 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.829562 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.829996 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.830658 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.831762 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.832544 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.833197 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.834270 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.835572 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.836305 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.837235 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.837891 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.838839 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.839757 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.840312 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.841282 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.842208 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.843249 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.844060 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.844962 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.845582 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.864685 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.884529 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.885285 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.885322 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.885333 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"93cc7f41778b13e5e73f961b7b8b5c99105d8c268c01c421e3321f52e5a98a12"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.886906 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e7553093737188a42b5e39fd29ade7f08b5c85102217839da550acf093dc070e"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.888706 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.888820 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.888906 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1862981857cab17f3f58c4debc8166ef2033b68028df52633246aaf371d144de"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.890983 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.892864 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.893305 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.893811 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hnf4q" event={"ID":"74b63880-331e-46a7-895b-434b52ff35b2","Type":"ContainerStarted","Data":"5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.893901 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hnf4q" event={"ID":"74b63880-331e-46a7-895b-434b52ff35b2","Type":"ContainerStarted","Data":"6190932be9cd604868254469c13e0c43a0c02422bfd1e30dc31124ea0f605cfc"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.895405 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerID="3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e" exitCode=0 Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.895520 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerDied","Data":"3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.895590 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerStarted","Data":"d9089e51b51e3ffe9659d4b0684eb789eb5eeaa74d610fc958e741c6c6637568"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.899233 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.899274 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"970b0e47a9543c3b2095fa664e6af456ad8d1db2afd231d8789cd4acc954afd3"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.902082 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.904588 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" event={"ID":"40784d67-bada-4229-8764-efba7c96edb5","Type":"ContainerStarted","Data":"039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.904744 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" event={"ID":"40784d67-bada-4229-8764-efba7c96edb5","Type":"ContainerStarted","Data":"b0f9d2fab03157ec021b17fb14e21ada9428d7a0636a552d2cc1a6ef15543f1e"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.905927 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xdxgd" event={"ID":"c6e55203-65ae-4767-9a96-f309bc41ddf3","Type":"ContainerStarted","Data":"6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.905973 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xdxgd" event={"ID":"c6e55203-65ae-4767-9a96-f309bc41ddf3","Type":"ContainerStarted","Data":"abdf024b4a7dfa923589012b4b6cd2999144df633c29e7445ead216194105b7f"} Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.928630 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.952767 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:39 crc kubenswrapper[4828]: I1210 18:55:39.968786 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.028333 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.056975 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.084596 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.100263 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.121985 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.137909 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.160582 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.177215 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.196155 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.214931 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.232919 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.246585 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.262862 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.287303 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.308162 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.322129 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.458019 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:40 crc kubenswrapper[4828]: E1210 18:55:40.458138 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:55:42.458117752 +0000 UTC m=+22.968728777 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.458189 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.458226 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:40 crc kubenswrapper[4828]: E1210 18:55:40.458343 4828 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:40 crc kubenswrapper[4828]: E1210 18:55:40.458349 4828 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:40 crc kubenswrapper[4828]: E1210 18:55:40.458385 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:42.45837634 +0000 UTC m=+22.968987355 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:40 crc kubenswrapper[4828]: E1210 18:55:40.458420 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:42.458403691 +0000 UTC m=+22.969014696 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.461584 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.493858 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.503514 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.559278 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.559395 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:40 crc kubenswrapper[4828]: E1210 18:55:40.559524 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:40 crc kubenswrapper[4828]: E1210 18:55:40.559548 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:40 crc kubenswrapper[4828]: E1210 18:55:40.559561 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:40 crc kubenswrapper[4828]: E1210 18:55:40.559575 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:40 crc kubenswrapper[4828]: E1210 18:55:40.559597 4828 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:40 crc kubenswrapper[4828]: E1210 18:55:40.559576 4828 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:40 crc kubenswrapper[4828]: E1210 18:55:40.559659 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:42.559638739 +0000 UTC m=+23.070249754 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:40 crc kubenswrapper[4828]: E1210 18:55:40.559712 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:42.55969224 +0000 UTC m=+23.070303315 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.605633 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.625722 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.628791 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.677887 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.725413 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.726350 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.788220 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.788268 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:40 crc kubenswrapper[4828]: E1210 18:55:40.788366 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.788399 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:40 crc kubenswrapper[4828]: E1210 18:55:40.788511 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:40 crc kubenswrapper[4828]: E1210 18:55:40.788612 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.814388 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.816476 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.828319 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-5px8v"] Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.828843 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5px8v" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.830772 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.831205 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.831531 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.832787 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.842258 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.849687 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.854094 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.863001 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84wpn\" (UniqueName: \"kubernetes.io/projected/ff6072ff-b4d5-48a1-9ee5-03458bcbd758-kube-api-access-84wpn\") pod \"node-ca-5px8v\" (UID: \"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\") " pod="openshift-image-registry/node-ca-5px8v" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.863082 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ff6072ff-b4d5-48a1-9ee5-03458bcbd758-serviceca\") pod \"node-ca-5px8v\" (UID: \"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\") " pod="openshift-image-registry/node-ca-5px8v" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.863159 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ff6072ff-b4d5-48a1-9ee5-03458bcbd758-host\") pod \"node-ca-5px8v\" (UID: \"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\") " pod="openshift-image-registry/node-ca-5px8v" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.864879 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.880252 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.890448 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.898670 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.905886 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.911687 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.914939 4828 generic.go:334] "Generic (PLEG): container finished" podID="40784d67-bada-4229-8764-efba7c96edb5" containerID="039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06" exitCode=0 Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.915046 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" event={"ID":"40784d67-bada-4229-8764-efba7c96edb5","Type":"ContainerDied","Data":"039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06"} Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.923089 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerStarted","Data":"6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275"} Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.923146 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerStarted","Data":"a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77"} Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.923184 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerStarted","Data":"0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7"} Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.923195 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerStarted","Data":"f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908"} Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.926429 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.943064 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.960765 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.965630 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ff6072ff-b4d5-48a1-9ee5-03458bcbd758-host\") pod \"node-ca-5px8v\" (UID: \"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\") " pod="openshift-image-registry/node-ca-5px8v" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.965719 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84wpn\" (UniqueName: \"kubernetes.io/projected/ff6072ff-b4d5-48a1-9ee5-03458bcbd758-kube-api-access-84wpn\") pod \"node-ca-5px8v\" (UID: \"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\") " pod="openshift-image-registry/node-ca-5px8v" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.965757 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ff6072ff-b4d5-48a1-9ee5-03458bcbd758-host\") pod \"node-ca-5px8v\" (UID: \"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\") " pod="openshift-image-registry/node-ca-5px8v" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.965788 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ff6072ff-b4d5-48a1-9ee5-03458bcbd758-serviceca\") pod \"node-ca-5px8v\" (UID: \"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\") " pod="openshift-image-registry/node-ca-5px8v" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.966969 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/ff6072ff-b4d5-48a1-9ee5-03458bcbd758-serviceca\") pod \"node-ca-5px8v\" (UID: \"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\") " pod="openshift-image-registry/node-ca-5px8v" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.971390 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 10 18:55:40 crc kubenswrapper[4828]: I1210 18:55:40.990938 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.033188 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.051629 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.090742 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.092426 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84wpn\" (UniqueName: \"kubernetes.io/projected/ff6072ff-b4d5-48a1-9ee5-03458bcbd758-kube-api-access-84wpn\") pod \"node-ca-5px8v\" (UID: \"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\") " pod="openshift-image-registry/node-ca-5px8v" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.120027 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.131244 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.150684 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.167340 4828 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.169083 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.169104 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.169113 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.169179 4828 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.170648 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.173661 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5px8v" Dec 10 18:55:41 crc kubenswrapper[4828]: W1210 18:55:41.185823 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff6072ff_b4d5_48a1_9ee5_03458bcbd758.slice/crio-a48d073d2666ab4eea14409dca2c1e71fe9fbc8ec2e270ff6885b92d21219e7a WatchSource:0}: Error finding container a48d073d2666ab4eea14409dca2c1e71fe9fbc8ec2e270ff6885b92d21219e7a: Status 404 returned error can't find the container with id a48d073d2666ab4eea14409dca2c1e71fe9fbc8ec2e270ff6885b92d21219e7a Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.192221 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.211137 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.251911 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.271063 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.294968 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.330549 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.358703 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.371454 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.390924 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.412106 4828 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.412339 4828 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.413608 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.413675 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.413690 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.413736 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.413748 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:41Z","lastTransitionTime":"2025-12-10T18:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:41 crc kubenswrapper[4828]: E1210 18:55:41.434026 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.440104 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.440151 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.440167 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.440189 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.440204 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:41Z","lastTransitionTime":"2025-12-10T18:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:41 crc kubenswrapper[4828]: E1210 18:55:41.453178 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.456406 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.456452 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.456466 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.456486 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.456497 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:41Z","lastTransitionTime":"2025-12-10T18:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.462630 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: E1210 18:55:41.468495 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.471555 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.471590 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.471599 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.471612 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.471622 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:41Z","lastTransitionTime":"2025-12-10T18:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:41 crc kubenswrapper[4828]: E1210 18:55:41.482816 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.485793 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.485837 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.485847 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.485860 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.485870 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:41Z","lastTransitionTime":"2025-12-10T18:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:41 crc kubenswrapper[4828]: E1210 18:55:41.498961 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: E1210 18:55:41.499072 4828 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.500489 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.500599 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.500660 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.500724 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.500781 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:41Z","lastTransitionTime":"2025-12-10T18:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.503387 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.539638 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.579599 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.602849 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.602879 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.602888 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.602901 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.602912 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:41Z","lastTransitionTime":"2025-12-10T18:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.622715 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.657649 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.699852 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.705686 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.705723 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.705737 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.705753 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.705766 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:41Z","lastTransitionTime":"2025-12-10T18:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.740651 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.778402 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.807697 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.807749 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.807761 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.807778 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.807789 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:41Z","lastTransitionTime":"2025-12-10T18:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.823301 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.858821 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.899683 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.910503 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.910545 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.910559 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.910576 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.910588 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:41Z","lastTransitionTime":"2025-12-10T18:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.929322 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerStarted","Data":"62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2"} Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.929371 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerStarted","Data":"279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e"} Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.931564 4828 generic.go:334] "Generic (PLEG): container finished" podID="40784d67-bada-4229-8764-efba7c96edb5" containerID="55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e" exitCode=0 Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.931624 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" event={"ID":"40784d67-bada-4229-8764-efba7c96edb5","Type":"ContainerDied","Data":"55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e"} Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.934196 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5px8v" event={"ID":"ff6072ff-b4d5-48a1-9ee5-03458bcbd758","Type":"ContainerStarted","Data":"95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073"} Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.934238 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5px8v" event={"ID":"ff6072ff-b4d5-48a1-9ee5-03458bcbd758","Type":"ContainerStarted","Data":"a48d073d2666ab4eea14409dca2c1e71fe9fbc8ec2e270ff6885b92d21219e7a"} Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.935748 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0"} Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.946176 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:41 crc kubenswrapper[4828]: I1210 18:55:41.984670 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.013335 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.013386 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.013396 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.013412 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.013423 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:42Z","lastTransitionTime":"2025-12-10T18:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.018560 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.060280 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.100269 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.115163 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.115212 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.115225 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.115243 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.115256 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:42Z","lastTransitionTime":"2025-12-10T18:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.138614 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.184959 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.217527 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.217569 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.217580 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.217595 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.217607 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:42Z","lastTransitionTime":"2025-12-10T18:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.219091 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.262313 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.324328 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.324375 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.324388 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.324404 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.324415 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:42Z","lastTransitionTime":"2025-12-10T18:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.328898 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.347300 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.377351 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.419602 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.426976 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.427022 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.427037 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.427057 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.427066 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:42Z","lastTransitionTime":"2025-12-10T18:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.459786 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.479092 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:42 crc kubenswrapper[4828]: E1210 18:55:42.479193 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:55:46.479173282 +0000 UTC m=+26.989784287 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.479249 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.479275 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:42 crc kubenswrapper[4828]: E1210 18:55:42.479349 4828 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:42 crc kubenswrapper[4828]: E1210 18:55:42.479375 4828 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:42 crc kubenswrapper[4828]: E1210 18:55:42.479380 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:46.479373669 +0000 UTC m=+26.989984674 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:42 crc kubenswrapper[4828]: E1210 18:55:42.479415 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:46.4794065 +0000 UTC m=+26.990017495 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.497256 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.529962 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.529993 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.530001 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.530014 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.530022 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:42Z","lastTransitionTime":"2025-12-10T18:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.538056 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.580355 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.580419 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:42 crc kubenswrapper[4828]: E1210 18:55:42.580522 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:42 crc kubenswrapper[4828]: E1210 18:55:42.580536 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:42 crc kubenswrapper[4828]: E1210 18:55:42.580579 4828 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:42 crc kubenswrapper[4828]: E1210 18:55:42.580624 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:46.580612156 +0000 UTC m=+27.091223161 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:42 crc kubenswrapper[4828]: E1210 18:55:42.580634 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:42 crc kubenswrapper[4828]: E1210 18:55:42.580673 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:42 crc kubenswrapper[4828]: E1210 18:55:42.580687 4828 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:42 crc kubenswrapper[4828]: E1210 18:55:42.580759 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:46.58073606 +0000 UTC m=+27.091347065 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.581674 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.622251 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.633294 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.633324 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.633336 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.633350 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.633359 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:42Z","lastTransitionTime":"2025-12-10T18:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.665731 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.700041 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.735866 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.735902 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.735914 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.735933 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.735945 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:42Z","lastTransitionTime":"2025-12-10T18:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.788519 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.788599 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:42 crc kubenswrapper[4828]: E1210 18:55:42.788656 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.788543 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:42 crc kubenswrapper[4828]: E1210 18:55:42.788859 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:42 crc kubenswrapper[4828]: E1210 18:55:42.788977 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.838333 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.838374 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.838402 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.838419 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.838429 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:42Z","lastTransitionTime":"2025-12-10T18:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.939602 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.939634 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.939642 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.939656 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.939667 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:42Z","lastTransitionTime":"2025-12-10T18:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.941868 4828 generic.go:334] "Generic (PLEG): container finished" podID="40784d67-bada-4229-8764-efba7c96edb5" containerID="acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf" exitCode=0 Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.941977 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" event={"ID":"40784d67-bada-4229-8764-efba7c96edb5","Type":"ContainerDied","Data":"acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf"} Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.958330 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.974613 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:42 crc kubenswrapper[4828]: I1210 18:55:42.990095 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.007344 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.016911 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.032986 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.042373 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.042413 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.042425 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.042442 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.042453 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:43Z","lastTransitionTime":"2025-12-10T18:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.045216 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.056789 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.068203 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.097559 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.141334 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.144834 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.144863 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.144873 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.144888 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.144899 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:43Z","lastTransitionTime":"2025-12-10T18:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.179790 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.222106 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.247276 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.247312 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.247319 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.247333 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.247344 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:43Z","lastTransitionTime":"2025-12-10T18:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.264779 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.349862 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.349900 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.349908 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.349924 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.349936 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:43Z","lastTransitionTime":"2025-12-10T18:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.452177 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.452209 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.452217 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.452230 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.452240 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:43Z","lastTransitionTime":"2025-12-10T18:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.554525 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.554565 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.554576 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.554590 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.554600 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:43Z","lastTransitionTime":"2025-12-10T18:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.657196 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.657258 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.657278 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.657302 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.657315 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:43Z","lastTransitionTime":"2025-12-10T18:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.760216 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.760285 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.760302 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.760329 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.760359 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:43Z","lastTransitionTime":"2025-12-10T18:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.864062 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.864133 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.864151 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.864175 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.864194 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:43Z","lastTransitionTime":"2025-12-10T18:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.950881 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerStarted","Data":"db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba"} Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.953616 4828 generic.go:334] "Generic (PLEG): container finished" podID="40784d67-bada-4229-8764-efba7c96edb5" containerID="9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2" exitCode=0 Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.953672 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" event={"ID":"40784d67-bada-4229-8764-efba7c96edb5","Type":"ContainerDied","Data":"9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2"} Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.968096 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.968128 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.968139 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.968156 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.968167 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:43Z","lastTransitionTime":"2025-12-10T18:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.970526 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:43 crc kubenswrapper[4828]: I1210 18:55:43.991602 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.004624 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.017324 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.039606 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.050704 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.063477 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.070469 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.070504 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.070515 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.070529 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.070539 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:44Z","lastTransitionTime":"2025-12-10T18:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.075667 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.089148 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.107379 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.130767 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.151254 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.166100 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.172241 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.172278 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.172291 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.172308 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.172319 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:44Z","lastTransitionTime":"2025-12-10T18:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.181167 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.274594 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.274637 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.274647 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.274662 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.274673 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:44Z","lastTransitionTime":"2025-12-10T18:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.376468 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.376505 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.376514 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.376528 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.376538 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:44Z","lastTransitionTime":"2025-12-10T18:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.479190 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.479224 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.479233 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.479245 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.479254 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:44Z","lastTransitionTime":"2025-12-10T18:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.581193 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.581231 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.581239 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.581252 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.581264 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:44Z","lastTransitionTime":"2025-12-10T18:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.683998 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.684040 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.684052 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.684067 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.684077 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:44Z","lastTransitionTime":"2025-12-10T18:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.785825 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.785870 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.785885 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.785900 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.785909 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:44Z","lastTransitionTime":"2025-12-10T18:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.788066 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.788066 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.788163 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:44 crc kubenswrapper[4828]: E1210 18:55:44.788263 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:44 crc kubenswrapper[4828]: E1210 18:55:44.788320 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:44 crc kubenswrapper[4828]: E1210 18:55:44.788425 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.888460 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.888517 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.888527 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.888541 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.888552 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:44Z","lastTransitionTime":"2025-12-10T18:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.961609 4828 generic.go:334] "Generic (PLEG): container finished" podID="40784d67-bada-4229-8764-efba7c96edb5" containerID="988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9" exitCode=0 Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.961674 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" event={"ID":"40784d67-bada-4229-8764-efba7c96edb5","Type":"ContainerDied","Data":"988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9"} Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.974467 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.987864 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.990516 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.990549 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.990563 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.990581 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.990595 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:44Z","lastTransitionTime":"2025-12-10T18:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:44 crc kubenswrapper[4828]: I1210 18:55:44.998370 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:44Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.011212 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.026904 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.044241 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.061424 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.074263 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.086517 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.093479 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.093529 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.093540 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.093559 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.093572 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:45Z","lastTransitionTime":"2025-12-10T18:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.105145 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.123196 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.135743 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.146350 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.158098 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.196293 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.196337 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.196347 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.196364 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.196377 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:45Z","lastTransitionTime":"2025-12-10T18:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.298365 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.298399 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.298408 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.298421 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.298431 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:45Z","lastTransitionTime":"2025-12-10T18:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.402515 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.402776 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.402788 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.402826 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.402838 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:45Z","lastTransitionTime":"2025-12-10T18:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.506206 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.506249 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.506274 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.506289 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.506300 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:45Z","lastTransitionTime":"2025-12-10T18:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.608609 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.608643 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.608654 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.608670 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.608682 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:45Z","lastTransitionTime":"2025-12-10T18:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.711205 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.711269 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.711291 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.711314 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.711329 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:45Z","lastTransitionTime":"2025-12-10T18:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.813186 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.813232 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.813249 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.813268 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.813283 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:45Z","lastTransitionTime":"2025-12-10T18:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.916220 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.916257 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.916269 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.916290 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.916305 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:45Z","lastTransitionTime":"2025-12-10T18:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.974293 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerStarted","Data":"4a0b8a8c3744910d9d70a4e24694f7e005fce4509a32e3c413cb47c928ac8dce"} Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.976144 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.976279 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.984332 4828 generic.go:334] "Generic (PLEG): container finished" podID="40784d67-bada-4229-8764-efba7c96edb5" containerID="a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58" exitCode=0 Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.984397 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" event={"ID":"40784d67-bada-4229-8764-efba7c96edb5","Type":"ContainerDied","Data":"a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58"} Dec 10 18:55:45 crc kubenswrapper[4828]: I1210 18:55:45.990317 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:45Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.007205 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.007314 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.010727 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.019720 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.019775 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.019832 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.019866 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.019891 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:46Z","lastTransitionTime":"2025-12-10T18:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.029900 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.052861 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.066179 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.081697 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.101620 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a0b8a8c3744910d9d70a4e24694f7e005fce4509a32e3c413cb47c928ac8dce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.114186 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.123357 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.123398 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.123409 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.123427 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.123440 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:46Z","lastTransitionTime":"2025-12-10T18:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.130531 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.143469 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.157495 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.171062 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.190494 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.205342 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.219503 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.225012 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.225040 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.225048 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.225065 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.225074 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:46Z","lastTransitionTime":"2025-12-10T18:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.235419 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.249498 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.264339 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.279038 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.291672 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.304757 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.328119 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.328417 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.328551 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.328683 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.328845 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:46Z","lastTransitionTime":"2025-12-10T18:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.341609 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.365003 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.381222 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.397291 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.412853 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a0b8a8c3744910d9d70a4e24694f7e005fce4509a32e3c413cb47c928ac8dce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.431513 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.431557 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.431568 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.431586 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.431597 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:46Z","lastTransitionTime":"2025-12-10T18:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.431828 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.445712 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:46Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.517532 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.517651 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:46 crc kubenswrapper[4828]: E1210 18:55:46.517689 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:55:54.51765834 +0000 UTC m=+35.028269365 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:55:46 crc kubenswrapper[4828]: E1210 18:55:46.517733 4828 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:46 crc kubenswrapper[4828]: E1210 18:55:46.517795 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:54.517778024 +0000 UTC m=+35.028389039 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.517837 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:46 crc kubenswrapper[4828]: E1210 18:55:46.517978 4828 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:46 crc kubenswrapper[4828]: E1210 18:55:46.518081 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:54.518059843 +0000 UTC m=+35.028670858 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.534422 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.534455 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.534464 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.534477 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.534486 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:46Z","lastTransitionTime":"2025-12-10T18:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.592774 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.619187 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.619237 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:46 crc kubenswrapper[4828]: E1210 18:55:46.619359 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:46 crc kubenswrapper[4828]: E1210 18:55:46.619377 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:46 crc kubenswrapper[4828]: E1210 18:55:46.619402 4828 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:46 crc kubenswrapper[4828]: E1210 18:55:46.619424 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:46 crc kubenswrapper[4828]: E1210 18:55:46.619463 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:46 crc kubenswrapper[4828]: E1210 18:55:46.619478 4828 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:46 crc kubenswrapper[4828]: E1210 18:55:46.619451 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:54.619437234 +0000 UTC m=+35.130048239 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:46 crc kubenswrapper[4828]: E1210 18:55:46.619561 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:54.619540058 +0000 UTC m=+35.130151083 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.637124 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.637182 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.637199 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.637220 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.637238 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:46Z","lastTransitionTime":"2025-12-10T18:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.739948 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.740005 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.740025 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.740052 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.740070 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:46Z","lastTransitionTime":"2025-12-10T18:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.788704 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.788697 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.788865 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:46 crc kubenswrapper[4828]: E1210 18:55:46.788849 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:46 crc kubenswrapper[4828]: E1210 18:55:46.789299 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:46 crc kubenswrapper[4828]: E1210 18:55:46.789478 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.842935 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.842984 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.842998 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.843019 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.843035 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:46Z","lastTransitionTime":"2025-12-10T18:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.945567 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.945622 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.945634 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.945652 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.945664 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:46Z","lastTransitionTime":"2025-12-10T18:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:46 crc kubenswrapper[4828]: I1210 18:55:46.994517 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" event={"ID":"40784d67-bada-4229-8764-efba7c96edb5","Type":"ContainerStarted","Data":"cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587"} Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.010435 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.021475 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.036402 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.048820 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.048864 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.048877 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.048895 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.048907 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:47Z","lastTransitionTime":"2025-12-10T18:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.049472 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.063900 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.077157 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.093519 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.114828 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a0b8a8c3744910d9d70a4e24694f7e005fce4509a32e3c413cb47c928ac8dce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.135016 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.147856 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.151878 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.151936 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.151953 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.151986 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.152003 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:47Z","lastTransitionTime":"2025-12-10T18:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.163987 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.175526 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.187879 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.198014 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:47Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.255046 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.255086 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.255094 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.255107 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.255116 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:47Z","lastTransitionTime":"2025-12-10T18:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.357475 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.357524 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.357544 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.357563 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.357576 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:47Z","lastTransitionTime":"2025-12-10T18:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.460105 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.460155 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.460165 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.460181 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.460192 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:47Z","lastTransitionTime":"2025-12-10T18:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.562502 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.562544 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.562553 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.562567 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.562580 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:47Z","lastTransitionTime":"2025-12-10T18:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.665059 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.665103 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.665113 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.665129 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.665141 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:47Z","lastTransitionTime":"2025-12-10T18:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.767526 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.767596 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.767611 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.767649 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.767663 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:47Z","lastTransitionTime":"2025-12-10T18:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.870503 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.870538 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.870545 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.870558 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.870568 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:47Z","lastTransitionTime":"2025-12-10T18:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.972850 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.972896 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.972912 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.972931 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:47 crc kubenswrapper[4828]: I1210 18:55:47.972943 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:47Z","lastTransitionTime":"2025-12-10T18:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.075978 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.076052 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.076061 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.076077 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.076087 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:48Z","lastTransitionTime":"2025-12-10T18:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.178854 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.178916 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.178935 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.178962 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.178980 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:48Z","lastTransitionTime":"2025-12-10T18:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.281847 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.281894 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.281905 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.281923 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.281933 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:48Z","lastTransitionTime":"2025-12-10T18:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.385295 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.385355 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.385368 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.385389 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.385402 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:48Z","lastTransitionTime":"2025-12-10T18:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.488559 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.488607 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.488617 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.488632 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.488644 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:48Z","lastTransitionTime":"2025-12-10T18:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.592303 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.592365 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.592378 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.592404 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.592421 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:48Z","lastTransitionTime":"2025-12-10T18:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.694882 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.694933 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.694943 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.694958 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.694968 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:48Z","lastTransitionTime":"2025-12-10T18:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.788131 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.788164 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.788309 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:48 crc kubenswrapper[4828]: E1210 18:55:48.788400 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:48 crc kubenswrapper[4828]: E1210 18:55:48.788565 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:48 crc kubenswrapper[4828]: E1210 18:55:48.788681 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.797957 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.798041 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.798053 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.798068 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.798077 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:48Z","lastTransitionTime":"2025-12-10T18:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.901467 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.902511 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.902580 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.902624 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:48 crc kubenswrapper[4828]: I1210 18:55:48.902652 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:48Z","lastTransitionTime":"2025-12-10T18:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.004341 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovnkube-controller/0.log" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.008627 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.008717 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.008742 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.008778 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.008844 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.008980 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerID="4a0b8a8c3744910d9d70a4e24694f7e005fce4509a32e3c413cb47c928ac8dce" exitCode=1 Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.009033 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerDied","Data":"4a0b8a8c3744910d9d70a4e24694f7e005fce4509a32e3c413cb47c928ac8dce"} Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.009921 4828 scope.go:117] "RemoveContainer" containerID="4a0b8a8c3744910d9d70a4e24694f7e005fce4509a32e3c413cb47c928ac8dce" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.032168 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.043393 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.060197 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.081199 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a0b8a8c3744910d9d70a4e24694f7e005fce4509a32e3c413cb47c928ac8dce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0b8a8c3744910d9d70a4e24694f7e005fce4509a32e3c413cb47c928ac8dce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"oval\\\\nI1210 18:55:48.089543 6109 handler.go:208] Removed *v1.Node event handler 7\\\\nI1210 18:55:48.089547 6109 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:55:48.089575 6109 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.089618 6109 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.089646 6109 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.089571 6109 handler.go:208] Removed *v1.Node event handler 2\\\\nI1210 18:55:48.089736 6109 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.090374 6109 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 18:55:48.090409 6109 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 18:55:48.090442 6109 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 18:55:48.090462 6109 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 18:55:48.090487 6109 factory.go:656] Stopping watch factory\\\\nI1210 18:55:48.090489 6109 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 18:55:48.090501 6109 ovnkube.go:599] Stopped ovnkube\\\\nI1210 18:55:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.093692 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.106072 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.112129 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.112221 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.112241 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.112275 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.112298 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.118594 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.132833 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.142764 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.155420 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.167519 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.180880 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.193261 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.203036 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.214447 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.214484 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.214494 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.214509 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.214520 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.317215 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.317255 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.317265 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.317278 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.317287 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.419537 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.419579 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.419592 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.419611 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.419622 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.521916 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.521964 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.521976 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.521991 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.522004 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.623580 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.623626 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.623635 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.623657 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.623668 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.725990 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.726932 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.726966 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.726990 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.727006 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.801358 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.810975 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.824409 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.829352 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.829386 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.829394 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.829412 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.829421 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.837485 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.848595 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.861033 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.873246 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.883700 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.895361 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.904486 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.915566 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.925473 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.932422 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.932461 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.932471 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.932525 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.932536 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:49Z","lastTransitionTime":"2025-12-10T18:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.942083 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:49 crc kubenswrapper[4828]: I1210 18:55:49.958340 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a0b8a8c3744910d9d70a4e24694f7e005fce4509a32e3c413cb47c928ac8dce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0b8a8c3744910d9d70a4e24694f7e005fce4509a32e3c413cb47c928ac8dce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"oval\\\\nI1210 18:55:48.089543 6109 handler.go:208] Removed *v1.Node event handler 7\\\\nI1210 18:55:48.089547 6109 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:55:48.089575 6109 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.089618 6109 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.089646 6109 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.089571 6109 handler.go:208] Removed *v1.Node event handler 2\\\\nI1210 18:55:48.089736 6109 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.090374 6109 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 18:55:48.090409 6109 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 18:55:48.090442 6109 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 18:55:48.090462 6109 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 18:55:48.090487 6109 factory.go:656] Stopping watch factory\\\\nI1210 18:55:48.090489 6109 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 18:55:48.090501 6109 ovnkube.go:599] Stopped ovnkube\\\\nI1210 18:55:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:49Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.012974 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovnkube-controller/0.log" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.015511 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerStarted","Data":"2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1"} Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.015958 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.027262 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.034724 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.034775 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.034786 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.034820 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.034834 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.038131 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.049082 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.058699 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.069233 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.076559 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.086948 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.096524 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.104845 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.114913 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.126737 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.136415 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.136463 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.136472 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.136486 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.136495 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.138501 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.151483 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.168438 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0b8a8c3744910d9d70a4e24694f7e005fce4509a32e3c413cb47c928ac8dce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"oval\\\\nI1210 18:55:48.089543 6109 handler.go:208] Removed *v1.Node event handler 7\\\\nI1210 18:55:48.089547 6109 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:55:48.089575 6109 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.089618 6109 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.089646 6109 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.089571 6109 handler.go:208] Removed *v1.Node event handler 2\\\\nI1210 18:55:48.089736 6109 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.090374 6109 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 18:55:48.090409 6109 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 18:55:48.090442 6109 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 18:55:48.090462 6109 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 18:55:48.090487 6109 factory.go:656] Stopping watch factory\\\\nI1210 18:55:48.090489 6109 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 18:55:48.090501 6109 ovnkube.go:599] Stopped ovnkube\\\\nI1210 18:55:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:50Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.238402 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.238445 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.238460 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.238480 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.238494 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.340129 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.340156 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.340166 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.340184 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.340194 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.442420 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.442773 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.443134 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.443433 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.443711 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.545887 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.546117 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.546192 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.546267 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.546330 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.648597 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.648870 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.648975 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.649079 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.649198 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.751565 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.751611 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.751624 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.751639 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.751648 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.788345 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.788317 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:50 crc kubenswrapper[4828]: E1210 18:55:50.788559 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.788368 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:50 crc kubenswrapper[4828]: E1210 18:55:50.788668 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:50 crc kubenswrapper[4828]: E1210 18:55:50.788702 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.853737 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.853771 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.853781 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.853818 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.853842 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.955755 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.955787 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.955813 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.955828 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:50 crc kubenswrapper[4828]: I1210 18:55:50.955836 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:50Z","lastTransitionTime":"2025-12-10T18:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.022402 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovnkube-controller/1.log" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.023085 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovnkube-controller/0.log" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.025624 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerID="2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1" exitCode=1 Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.025663 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerDied","Data":"2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1"} Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.025715 4828 scope.go:117] "RemoveContainer" containerID="4a0b8a8c3744910d9d70a4e24694f7e005fce4509a32e3c413cb47c928ac8dce" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.026352 4828 scope.go:117] "RemoveContainer" containerID="2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1" Dec 10 18:55:51 crc kubenswrapper[4828]: E1210 18:55:51.026504 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.042241 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.053772 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.057845 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.057873 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.057881 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.057898 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.057908 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.064930 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.076256 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.088770 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.102172 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.116637 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.134817 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.153554 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0b8a8c3744910d9d70a4e24694f7e005fce4509a32e3c413cb47c928ac8dce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"oval\\\\nI1210 18:55:48.089543 6109 handler.go:208] Removed *v1.Node event handler 7\\\\nI1210 18:55:48.089547 6109 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:55:48.089575 6109 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.089618 6109 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.089646 6109 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.089571 6109 handler.go:208] Removed *v1.Node event handler 2\\\\nI1210 18:55:48.089736 6109 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.090374 6109 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 18:55:48.090409 6109 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 18:55:48.090442 6109 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 18:55:48.090462 6109 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 18:55:48.090487 6109 factory.go:656] Stopping watch factory\\\\nI1210 18:55:48.090489 6109 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 18:55:48.090501 6109 ovnkube.go:599] Stopped ovnkube\\\\nI1210 18:55:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"message\\\":\\\":Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:49.782699 6242 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 18:55:49.782751 6242 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.160227 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.160253 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.160262 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.160277 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.160286 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.169862 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.180224 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.190484 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.202122 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.210549 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.262935 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.262982 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.262991 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.263008 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.263020 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.365016 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.365084 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.365102 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.365118 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.365128 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.467429 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.467478 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.467488 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.467502 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.467512 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.569405 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.569454 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.569467 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.569495 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.569509 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.632195 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.632233 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.632241 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.632257 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.632267 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4828]: E1210 18:55:51.642577 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.645899 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.645942 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.645954 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.645975 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.645987 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4828]: E1210 18:55:51.657415 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.660565 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.660669 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.660690 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.660716 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.660733 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4828]: E1210 18:55:51.677616 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.681150 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.681187 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.681197 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.681213 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.681223 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4828]: E1210 18:55:51.693705 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.697721 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.697777 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.697792 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.697835 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.697850 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4828]: E1210 18:55:51.709918 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: E1210 18:55:51.710038 4828 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.711385 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.711439 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.711452 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.711471 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.711484 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.813987 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.814045 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.814056 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.814070 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.814080 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.917194 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.917233 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.917243 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.917258 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.917269 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:51Z","lastTransitionTime":"2025-12-10T18:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.930162 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml"] Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.930599 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.933724 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.936496 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.951775 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.965520 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.973849 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fb4e3e43-3654-4f02-9ef7-d9ab9e666da2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pwjml\" (UID: \"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.974101 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fb4e3e43-3654-4f02-9ef7-d9ab9e666da2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pwjml\" (UID: \"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.974236 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw5sr\" (UniqueName: \"kubernetes.io/projected/fb4e3e43-3654-4f02-9ef7-d9ab9e666da2-kube-api-access-kw5sr\") pod \"ovnkube-control-plane-749d76644c-pwjml\" (UID: \"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.974271 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fb4e3e43-3654-4f02-9ef7-d9ab9e666da2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pwjml\" (UID: \"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" Dec 10 18:55:51 crc kubenswrapper[4828]: I1210 18:55:51.999864 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:51Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.017488 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.019407 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.019548 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.019648 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.019736 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.019812 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.030173 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovnkube-controller/1.log" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.031276 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.033465 4828 scope.go:117] "RemoveContainer" containerID="2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1" Dec 10 18:55:52 crc kubenswrapper[4828]: E1210 18:55:52.033673 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.043406 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.054340 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.062155 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.070899 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.074574 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fb4e3e43-3654-4f02-9ef7-d9ab9e666da2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pwjml\" (UID: \"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.074720 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw5sr\" (UniqueName: \"kubernetes.io/projected/fb4e3e43-3654-4f02-9ef7-d9ab9e666da2-kube-api-access-kw5sr\") pod \"ovnkube-control-plane-749d76644c-pwjml\" (UID: \"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.074849 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fb4e3e43-3654-4f02-9ef7-d9ab9e666da2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pwjml\" (UID: \"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.074936 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fb4e3e43-3654-4f02-9ef7-d9ab9e666da2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pwjml\" (UID: \"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.075339 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fb4e3e43-3654-4f02-9ef7-d9ab9e666da2-env-overrides\") pod \"ovnkube-control-plane-749d76644c-pwjml\" (UID: \"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.075658 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fb4e3e43-3654-4f02-9ef7-d9ab9e666da2-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-pwjml\" (UID: \"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.082617 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fb4e3e43-3654-4f02-9ef7-d9ab9e666da2-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-pwjml\" (UID: \"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.086715 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.092319 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw5sr\" (UniqueName: \"kubernetes.io/projected/fb4e3e43-3654-4f02-9ef7-d9ab9e666da2-kube-api-access-kw5sr\") pod \"ovnkube-control-plane-749d76644c-pwjml\" (UID: \"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.097552 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.110849 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.121534 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.121568 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.121577 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.121593 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.121603 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.131704 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4a0b8a8c3744910d9d70a4e24694f7e005fce4509a32e3c413cb47c928ac8dce\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:48Z\\\",\\\"message\\\":\\\"oval\\\\nI1210 18:55:48.089543 6109 handler.go:208] Removed *v1.Node event handler 7\\\\nI1210 18:55:48.089547 6109 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1210 18:55:48.089575 6109 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.089618 6109 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.089646 6109 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.089571 6109 handler.go:208] Removed *v1.Node event handler 2\\\\nI1210 18:55:48.089736 6109 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1210 18:55:48.090374 6109 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1210 18:55:48.090409 6109 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1210 18:55:48.090442 6109 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1210 18:55:48.090462 6109 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1210 18:55:48.090487 6109 factory.go:656] Stopping watch factory\\\\nI1210 18:55:48.090489 6109 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1210 18:55:48.090501 6109 ovnkube.go:599] Stopped ovnkube\\\\nI1210 18:55:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"message\\\":\\\":Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:49.782699 6242 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 18:55:49.782751 6242 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.144732 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.155587 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.167874 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.180203 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.204978 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.222001 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"message\\\":\\\":Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:49.782699 6242 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 18:55:49.782751 6242 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.223203 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.223231 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.223242 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.223255 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.223264 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.234293 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.243813 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.250626 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: W1210 18:55:52.257067 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb4e3e43_3654_4f02_9ef7_d9ab9e666da2.slice/crio-f138ffdefdedd7e6ce79b40d4dbcc462337c89ccad7a2ac2e0d4e7236745293e WatchSource:0}: Error finding container f138ffdefdedd7e6ce79b40d4dbcc462337c89ccad7a2ac2e0d4e7236745293e: Status 404 returned error can't find the container with id f138ffdefdedd7e6ce79b40d4dbcc462337c89ccad7a2ac2e0d4e7236745293e Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.262021 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.272917 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.286703 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.299184 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.309246 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.321256 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.325380 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.325412 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.325421 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.325436 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.325449 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.332182 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.341154 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.350609 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:52Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.427699 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.427729 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.427737 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.427749 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.427758 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.529773 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.529877 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.529901 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.529930 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.529954 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.632488 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.632554 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.632573 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.632598 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.632615 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.735487 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.735534 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.735546 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.735562 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.735573 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.788508 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.788553 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.788562 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:52 crc kubenswrapper[4828]: E1210 18:55:52.788650 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:52 crc kubenswrapper[4828]: E1210 18:55:52.788742 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:52 crc kubenswrapper[4828]: E1210 18:55:52.788903 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.837824 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.837854 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.837863 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.837878 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.837887 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.939722 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.939769 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.939778 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.939792 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:52 crc kubenswrapper[4828]: I1210 18:55:52.939817 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:52Z","lastTransitionTime":"2025-12-10T18:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.014510 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-74lmx"] Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.015095 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:55:53 crc kubenswrapper[4828]: E1210 18:55:53.015165 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.028332 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.037468 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" event={"ID":"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2","Type":"ContainerStarted","Data":"41e994b69b7552394d02f9225cf931743737e1bfcd0448aff5f860ab27afa00b"} Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.037514 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" event={"ID":"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2","Type":"ContainerStarted","Data":"e3f3052a777d2db499f91f40a7517c0f2cd48191aaf6ebc8b17f1c114c2623fc"} Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.037524 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" event={"ID":"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2","Type":"ContainerStarted","Data":"f138ffdefdedd7e6ce79b40d4dbcc462337c89ccad7a2ac2e0d4e7236745293e"} Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.041817 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.042449 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.042492 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.042502 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.042516 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.042526 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.051258 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.063085 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.075448 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.083813 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zq55\" (UniqueName: \"kubernetes.io/projected/472ff1d2-0936-4ee0-b738-dc26e2c03559-kube-api-access-6zq55\") pod \"network-metrics-daemon-74lmx\" (UID: \"472ff1d2-0936-4ee0-b738-dc26e2c03559\") " pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.083875 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs\") pod \"network-metrics-daemon-74lmx\" (UID: \"472ff1d2-0936-4ee0-b738-dc26e2c03559\") " pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.090657 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.102357 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.116908 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.129408 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.142981 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.145525 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.145580 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.145598 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.145622 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.145639 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.159322 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.173443 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-74lmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"472ff1d2-0936-4ee0-b738-dc26e2c03559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-74lmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.184547 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zq55\" (UniqueName: \"kubernetes.io/projected/472ff1d2-0936-4ee0-b738-dc26e2c03559-kube-api-access-6zq55\") pod \"network-metrics-daemon-74lmx\" (UID: \"472ff1d2-0936-4ee0-b738-dc26e2c03559\") " pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.184622 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs\") pod \"network-metrics-daemon-74lmx\" (UID: \"472ff1d2-0936-4ee0-b738-dc26e2c03559\") " pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:55:53 crc kubenswrapper[4828]: E1210 18:55:53.184942 4828 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:55:53 crc kubenswrapper[4828]: E1210 18:55:53.185081 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs podName:472ff1d2-0936-4ee0-b738-dc26e2c03559 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:53.685051167 +0000 UTC m=+34.195662212 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs") pod "network-metrics-daemon-74lmx" (UID: "472ff1d2-0936-4ee0-b738-dc26e2c03559") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.195698 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.205890 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zq55\" (UniqueName: \"kubernetes.io/projected/472ff1d2-0936-4ee0-b738-dc26e2c03559-kube-api-access-6zq55\") pod \"network-metrics-daemon-74lmx\" (UID: \"472ff1d2-0936-4ee0-b738-dc26e2c03559\") " pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.213711 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.234637 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.249274 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.249324 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.249338 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.249359 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.249392 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.261934 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"message\\\":\\\":Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:49.782699 6242 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 18:55:49.782751 6242 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.279972 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.298870 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.317840 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.343531 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"message\\\":\\\":Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:49.782699 6242 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 18:55:49.782751 6242 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.351918 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.351957 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.351968 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.351985 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.351997 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.363755 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-74lmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"472ff1d2-0936-4ee0-b738-dc26e2c03559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-74lmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.381696 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.399518 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.413883 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.430993 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.445349 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.455463 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.455549 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.455574 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.455609 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.455634 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.461471 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.480569 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.494573 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.509562 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.523288 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.540491 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f3052a777d2db499f91f40a7517c0f2cd48191aaf6ebc8b17f1c114c2623fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41e994b69b7552394d02f9225cf931743737e1bfcd0448aff5f860ab27afa00b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:53Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.558993 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.559060 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.559080 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.559106 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.559124 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.662591 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.662658 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.662677 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.662702 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.662722 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.691277 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs\") pod \"network-metrics-daemon-74lmx\" (UID: \"472ff1d2-0936-4ee0-b738-dc26e2c03559\") " pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:55:53 crc kubenswrapper[4828]: E1210 18:55:53.691513 4828 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:55:53 crc kubenswrapper[4828]: E1210 18:55:53.691608 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs podName:472ff1d2-0936-4ee0-b738-dc26e2c03559 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:54.691583425 +0000 UTC m=+35.202194470 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs") pod "network-metrics-daemon-74lmx" (UID: "472ff1d2-0936-4ee0-b738-dc26e2c03559") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.765385 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.765463 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.765481 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.765506 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.765523 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.868852 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.869340 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.869364 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.869391 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.869410 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.973170 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.973486 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.973590 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.973683 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:53 crc kubenswrapper[4828]: I1210 18:55:53.973768 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:53Z","lastTransitionTime":"2025-12-10T18:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.076556 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.076659 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.076678 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.076714 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.076733 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.179844 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.179916 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.179937 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.179959 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.179978 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.283785 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.283904 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.283926 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.283956 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.284008 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.387123 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.387165 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.387176 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.387191 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.387203 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.490688 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.490745 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.490755 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.490773 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.490786 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.593033 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.593098 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.593121 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.593149 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.593170 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.601958 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.602126 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.602237 4828 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.602313 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:56:10.602265338 +0000 UTC m=+51.112876373 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.602372 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:10.602352251 +0000 UTC m=+51.112963416 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.602433 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.602624 4828 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.602732 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:10.602701532 +0000 UTC m=+51.113312597 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.696212 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.696273 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.696291 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.696341 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.696359 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.704070 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.704155 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.704241 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs\") pod \"network-metrics-daemon-74lmx\" (UID: \"472ff1d2-0936-4ee0-b738-dc26e2c03559\") " pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.704394 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.704430 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.704450 4828 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.704460 4828 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.704540 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:10.704515448 +0000 UTC m=+51.215126493 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.704571 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs podName:472ff1d2-0936-4ee0-b738-dc26e2c03559 nodeName:}" failed. No retries permitted until 2025-12-10 18:55:56.704556689 +0000 UTC m=+37.215167734 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs") pod "network-metrics-daemon-74lmx" (UID: "472ff1d2-0936-4ee0-b738-dc26e2c03559") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.704914 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.704942 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.704954 4828 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.704997 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:10.704985623 +0000 UTC m=+51.215596708 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.788144 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.788218 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.788218 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.788313 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.788413 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.788569 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.788773 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:55:54 crc kubenswrapper[4828]: E1210 18:55:54.788951 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.799067 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.799160 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.799184 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.799217 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.799239 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.901411 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.901466 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.901474 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.901487 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:54 crc kubenswrapper[4828]: I1210 18:55:54.901499 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:54Z","lastTransitionTime":"2025-12-10T18:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.004696 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.004748 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.004762 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.004781 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.004793 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.069821 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.089501 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.108372 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.108420 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.108437 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.108457 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.108473 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.109644 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.132578 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.165996 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"message\\\":\\\":Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:49.782699 6242 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 18:55:49.782751 6242 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.184340 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-74lmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"472ff1d2-0936-4ee0-b738-dc26e2c03559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-74lmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.199519 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.211272 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.211337 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.211357 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.211379 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.211396 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.218751 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.239465 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.259883 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.274632 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.290680 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.310941 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.313706 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.313760 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.313777 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.313833 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.313853 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.330131 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.353230 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.366875 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.380989 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f3052a777d2db499f91f40a7517c0f2cd48191aaf6ebc8b17f1c114c2623fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41e994b69b7552394d02f9225cf931743737e1bfcd0448aff5f860ab27afa00b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:55Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.417089 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.417141 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.417154 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.417172 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.417184 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.520892 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.520960 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.520985 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.521020 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.521044 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.623335 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.623409 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.623427 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.623449 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.623464 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.727099 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.727194 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.727215 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.727243 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.727269 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.830333 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.830431 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.830452 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.830509 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.830530 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.933988 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.934048 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.934064 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.934090 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:55 crc kubenswrapper[4828]: I1210 18:55:55.934108 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:55Z","lastTransitionTime":"2025-12-10T18:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.037748 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.037793 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.037828 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.037849 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.037864 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.140356 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.140406 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.140423 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.140447 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.140465 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.242879 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.242972 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.242997 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.243034 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.243051 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.345240 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.345317 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.345330 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.345347 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.345360 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.448530 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.448597 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.448614 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.448639 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.448657 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.551883 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.551978 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.551991 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.552012 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.552025 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.655250 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.655314 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.655332 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.655358 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.655376 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.728263 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs\") pod \"network-metrics-daemon-74lmx\" (UID: \"472ff1d2-0936-4ee0-b738-dc26e2c03559\") " pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:55:56 crc kubenswrapper[4828]: E1210 18:55:56.728485 4828 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:55:56 crc kubenswrapper[4828]: E1210 18:55:56.728618 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs podName:472ff1d2-0936-4ee0-b738-dc26e2c03559 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:00.728585426 +0000 UTC m=+41.239196461 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs") pod "network-metrics-daemon-74lmx" (UID: "472ff1d2-0936-4ee0-b738-dc26e2c03559") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.759394 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.759474 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.759500 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.759531 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.759557 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.788050 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.788080 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:56 crc kubenswrapper[4828]: E1210 18:55:56.788539 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.788162 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:55:56 crc kubenswrapper[4828]: E1210 18:55:56.788680 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.788156 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:56 crc kubenswrapper[4828]: E1210 18:55:56.788751 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:55:56 crc kubenswrapper[4828]: E1210 18:55:56.788964 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.862930 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.863001 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.863024 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.863056 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.863080 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.965698 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.965741 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.965750 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.965764 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:56 crc kubenswrapper[4828]: I1210 18:55:56.965773 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:56Z","lastTransitionTime":"2025-12-10T18:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.068833 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.068874 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.068886 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.068905 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.068917 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.172318 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.172417 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.172456 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.172483 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.172501 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.276003 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.276072 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.276090 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.276117 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.276133 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.379343 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.379394 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.379420 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.379451 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.379473 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.482033 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.482087 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.482100 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.482120 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.482135 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.584928 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.584990 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.585009 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.585034 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.585056 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.687760 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.687838 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.687851 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.687871 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.687884 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.790434 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.790519 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.790530 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.790545 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.790559 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.893412 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.893473 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.893520 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.893542 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.893560 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.995788 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.995916 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.995937 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.995966 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:57 crc kubenswrapper[4828]: I1210 18:55:57.995991 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:57Z","lastTransitionTime":"2025-12-10T18:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.099251 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.099323 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.099461 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.099536 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.099559 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.202602 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.202664 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.202695 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.202722 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.202741 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.305841 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.305995 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.306024 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.306054 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.306075 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.408886 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.408952 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.408974 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.409007 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.409028 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.512978 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.513084 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.513106 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.513138 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.513161 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.616903 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.616960 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.617077 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.617146 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.617165 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.719446 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.719511 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.719532 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.719559 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.719579 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.787692 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.787752 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.787747 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:55:58 crc kubenswrapper[4828]: E1210 18:55:58.787846 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.787879 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:55:58 crc kubenswrapper[4828]: E1210 18:55:58.787939 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:55:58 crc kubenswrapper[4828]: E1210 18:55:58.788184 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:55:58 crc kubenswrapper[4828]: E1210 18:55:58.788166 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.821380 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.821431 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.821443 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.821461 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.821473 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.924300 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.924343 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.924355 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.924370 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:58 crc kubenswrapper[4828]: I1210 18:55:58.924380 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:58Z","lastTransitionTime":"2025-12-10T18:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.027132 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.027209 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.027217 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.027232 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.027241 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.129658 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.129732 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.129749 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.129773 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.129790 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.232202 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.232274 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.232297 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.232328 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.232354 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.335364 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.335495 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.335573 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.335599 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.335631 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.438716 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.438842 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.438870 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.438899 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.438921 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.542058 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.542112 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.542129 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.542151 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.542167 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.645713 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.645766 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.645779 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.645838 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.645850 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.748657 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.748711 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.748724 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.748743 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.748756 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.806782 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"message\\\":\\\":Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:49.782699 6242 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 18:55:49.782751 6242 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.818597 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-74lmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"472ff1d2-0936-4ee0-b738-dc26e2c03559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-74lmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.837990 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.850583 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.850618 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.850629 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.850646 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.850657 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.854022 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.871229 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.885925 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.899726 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.911356 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.920213 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.930500 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.940345 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.949693 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.952192 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.952220 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.952231 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.952246 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.952257 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:55:59Z","lastTransitionTime":"2025-12-10T18:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.959812 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f3052a777d2db499f91f40a7517c0f2cd48191aaf6ebc8b17f1c114c2623fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41e994b69b7552394d02f9225cf931743737e1bfcd0448aff5f860ab27afa00b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.977155 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:55:59 crc kubenswrapper[4828]: I1210 18:55:59.993135 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:55:59Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.004010 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:00Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.054644 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.054692 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.054703 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.054720 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.054731 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.157279 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.157306 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.157315 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.157328 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.157336 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.260965 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.261032 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.261051 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.261075 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.261093 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.364397 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.364456 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.364478 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.364507 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.364530 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.468219 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.468330 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.468356 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.468389 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.468415 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.570491 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.570550 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.570562 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.570579 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.570615 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.673463 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.673494 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.673503 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.673516 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.673542 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.766249 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs\") pod \"network-metrics-daemon-74lmx\" (UID: \"472ff1d2-0936-4ee0-b738-dc26e2c03559\") " pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:00 crc kubenswrapper[4828]: E1210 18:56:00.766567 4828 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:00 crc kubenswrapper[4828]: E1210 18:56:00.766685 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs podName:472ff1d2-0936-4ee0-b738-dc26e2c03559 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:08.766652779 +0000 UTC m=+49.277263814 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs") pod "network-metrics-daemon-74lmx" (UID: "472ff1d2-0936-4ee0-b738-dc26e2c03559") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.776199 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.776255 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.776268 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.776289 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.776303 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.788539 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.788547 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.788550 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.788692 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:00 crc kubenswrapper[4828]: E1210 18:56:00.788942 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:00 crc kubenswrapper[4828]: E1210 18:56:00.789046 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:00 crc kubenswrapper[4828]: E1210 18:56:00.789246 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:00 crc kubenswrapper[4828]: E1210 18:56:00.789356 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.879646 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.879708 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.879727 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.879757 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.879774 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.982479 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.982509 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.982516 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.982530 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:00 crc kubenswrapper[4828]: I1210 18:56:00.982539 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:00Z","lastTransitionTime":"2025-12-10T18:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.085445 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.085506 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.085533 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.085558 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.085575 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.187951 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.188020 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.188040 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.188066 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.188084 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.291595 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.291672 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.291691 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.291718 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.291754 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.395331 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.395388 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.395404 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.395433 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.395451 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.498462 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.498527 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.498551 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.498581 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.498603 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.601648 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.601728 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.601746 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.601766 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.601780 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.704850 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.704920 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.704942 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.704967 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.704985 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.808061 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.808128 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.808145 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.808170 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.808187 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.911678 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.911725 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.911738 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.911755 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.911766 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.955838 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.955914 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.955934 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.955965 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.955987 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4828]: E1210 18:56:01.969625 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.974205 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.974248 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.974260 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.974282 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:01 crc kubenswrapper[4828]: I1210 18:56:01.974295 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:01Z","lastTransitionTime":"2025-12-10T18:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:01 crc kubenswrapper[4828]: E1210 18:56:01.993414 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:01Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.000237 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.000352 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.000364 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.000381 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.000397 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4828]: E1210 18:56:02.014864 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.019589 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.019623 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.019631 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.019662 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.019672 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4828]: E1210 18:56:02.034200 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.038636 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.038687 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.038704 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.038726 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.038743 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4828]: E1210 18:56:02.051261 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:02Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:02 crc kubenswrapper[4828]: E1210 18:56:02.051495 4828 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.053216 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.053264 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.053277 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.053293 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.053304 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.155669 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.155704 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.155713 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.155728 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.155741 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.258977 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.259052 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.259075 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.259104 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.259125 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.363000 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.363073 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.363090 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.363116 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.363133 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.465745 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.465854 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.465875 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.465945 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.465966 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.568461 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.568539 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.568561 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.568590 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.568614 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.671688 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.672419 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.672454 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.672475 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.672488 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.774377 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.774439 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.774452 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.774471 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.774483 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.787874 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.787895 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.787943 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:02 crc kubenswrapper[4828]: E1210 18:56:02.788004 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.788032 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:02 crc kubenswrapper[4828]: E1210 18:56:02.788194 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:02 crc kubenswrapper[4828]: E1210 18:56:02.788340 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:02 crc kubenswrapper[4828]: E1210 18:56:02.788438 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.876967 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.877010 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.877021 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.877037 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.877049 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.979622 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.979702 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.979723 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.979754 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:02 crc kubenswrapper[4828]: I1210 18:56:02.979778 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:02Z","lastTransitionTime":"2025-12-10T18:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.083005 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.083074 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.083096 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.083125 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.083148 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.186621 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.186704 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.186727 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.186756 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.186776 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.290321 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.290397 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.290411 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.290427 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.290441 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.392669 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.392728 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.392742 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.392762 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.392775 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.495870 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.495913 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.495924 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.495941 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.495953 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.598063 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.598147 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.598169 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.598194 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.598211 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.701360 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.701431 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.701453 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.701482 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.701503 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.804146 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.804209 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.804228 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.804250 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.804267 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.906521 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.906562 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.906574 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.906590 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:03 crc kubenswrapper[4828]: I1210 18:56:03.906601 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:03Z","lastTransitionTime":"2025-12-10T18:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.010173 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.010234 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.010251 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.010275 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.010295 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.113486 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.113538 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.113556 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.113579 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.113599 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.216586 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.216643 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.216660 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.216681 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.216698 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.320145 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.320227 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.320251 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.320286 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.320301 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.423410 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.423517 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.423537 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.423560 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.423577 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.526699 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.526733 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.526752 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.526775 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.526792 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.629952 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.630008 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.630031 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.630058 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.630082 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.732756 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.732851 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.732876 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.732905 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.732927 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.788086 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.788256 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:04 crc kubenswrapper[4828]: E1210 18:56:04.788337 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.788135 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.788138 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:04 crc kubenswrapper[4828]: E1210 18:56:04.788445 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:04 crc kubenswrapper[4828]: E1210 18:56:04.788532 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:04 crc kubenswrapper[4828]: E1210 18:56:04.788680 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.836270 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.836311 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.836321 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.836337 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.836350 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.939182 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.939264 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.939288 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.939321 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:04 crc kubenswrapper[4828]: I1210 18:56:04.939346 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:04Z","lastTransitionTime":"2025-12-10T18:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.041833 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.041895 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.041910 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.041933 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.041949 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.144697 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.144749 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.144759 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.144775 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.144808 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.247366 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.247441 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.247466 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.247495 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.247517 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.350528 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.350608 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.350631 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.350662 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.350680 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.453950 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.454021 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.454053 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.454101 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.454128 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.556909 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.556964 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.556982 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.557014 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.557033 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.659636 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.659716 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.659737 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.659757 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.659771 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.762452 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.762522 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.762547 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.762577 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.762598 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.791048 4828 scope.go:117] "RemoveContainer" containerID="2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.865507 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.865906 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.866039 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.866192 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.866314 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.970191 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.970665 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.970681 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.970731 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:05 crc kubenswrapper[4828]: I1210 18:56:05.970749 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:05Z","lastTransitionTime":"2025-12-10T18:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.073676 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.073717 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.073725 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.073739 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.073749 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.086535 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovnkube-controller/1.log" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.091963 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerStarted","Data":"fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00"} Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.092355 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.113709 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.134060 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.157464 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.175622 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.175679 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.175688 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.175702 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.175711 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.177053 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.189731 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.209602 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.231018 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.244955 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.260874 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.273205 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.277972 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.278004 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.278030 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.278044 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.278052 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.284764 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f3052a777d2db499f91f40a7517c0f2cd48191aaf6ebc8b17f1c114c2623fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41e994b69b7552394d02f9225cf931743737e1bfcd0448aff5f860ab27afa00b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.296349 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.308001 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.322254 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.339345 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"message\\\":\\\":Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:49.782699 6242 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 18:55:49.782751 6242 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.349016 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-74lmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"472ff1d2-0936-4ee0-b738-dc26e2c03559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-74lmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.380822 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.380865 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.380876 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.380892 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.380905 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.482999 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.483034 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.483042 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.483055 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.483066 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.585383 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.585426 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.585437 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.585455 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.585467 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.688196 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.688231 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.688243 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.688261 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.688273 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.788171 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.788222 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:06 crc kubenswrapper[4828]: E1210 18:56:06.788399 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.788488 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.788499 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:06 crc kubenswrapper[4828]: E1210 18:56:06.788615 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:06 crc kubenswrapper[4828]: E1210 18:56:06.788887 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:06 crc kubenswrapper[4828]: E1210 18:56:06.789059 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.790398 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.790446 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.790470 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.790506 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.790529 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.893724 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.893777 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.893789 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.893846 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.893869 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.996512 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.996574 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.996590 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.996614 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:06 crc kubenswrapper[4828]: I1210 18:56:06.996631 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:06Z","lastTransitionTime":"2025-12-10T18:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.099606 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.099658 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.099674 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.099695 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.099712 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.100469 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovnkube-controller/2.log" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.101452 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovnkube-controller/1.log" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.106125 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerID="fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00" exitCode=1 Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.106166 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerDied","Data":"fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00"} Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.106200 4828 scope.go:117] "RemoveContainer" containerID="2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.106873 4828 scope.go:117] "RemoveContainer" containerID="fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00" Dec 10 18:56:07 crc kubenswrapper[4828]: E1210 18:56:07.107057 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.140974 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.165872 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.187463 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.202130 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.202165 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.202175 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.202191 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.202200 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.203357 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.218579 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.234323 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.245901 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.257986 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.270757 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.281273 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.291362 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f3052a777d2db499f91f40a7517c0f2cd48191aaf6ebc8b17f1c114c2623fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41e994b69b7552394d02f9225cf931743737e1bfcd0448aff5f860ab27afa00b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.305028 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.305097 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.305114 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.305138 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.305156 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.306512 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.319567 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.341756 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.360727 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2b10a2a82c01bba9532ffab3edfe69d3a6fdfb4f62f4ebfea5d000be1aadaab1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:55:50Z\\\",\\\"message\\\":\\\":Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1210 18:55:49.782699 6242 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.168:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {63b1440a-0908-4cab-8799-012fa1cf0b07}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1210 18:55:49.782751 6242 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:06Z\\\",\\\"message\\\":\\\"mplate:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.153\\\\\\\", Port:5443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1210 18:56:06.639052 6459 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z]\\\\nI1210 18:56:06.638925 6459 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-scheduler-operator for network=default : 3.861754ms\\\\nI1210 18:56:06.639267 6459 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.372716 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-74lmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"472ff1d2-0936-4ee0-b738-dc26e2c03559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-74lmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:07Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.407246 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.407298 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.407309 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.407327 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.407339 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.509527 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.509615 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.509648 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.509678 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.509699 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.612055 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.612124 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.612143 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.612167 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.612185 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.719770 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.719822 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.719832 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.719865 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.719874 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.822839 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.822870 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.822886 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.822901 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.822912 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.925338 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.925366 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.925374 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.925387 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:07 crc kubenswrapper[4828]: I1210 18:56:07.925395 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:07Z","lastTransitionTime":"2025-12-10T18:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.029164 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.029246 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.029264 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.029295 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.029315 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.112369 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovnkube-controller/2.log" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.117267 4828 scope.go:117] "RemoveContainer" containerID="fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00" Dec 10 18:56:08 crc kubenswrapper[4828]: E1210 18:56:08.117522 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.132128 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.132186 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.132205 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.132233 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.132253 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.132323 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.151437 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.164492 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.177264 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.192944 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.206953 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.220158 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.234475 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.234552 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.234570 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.234591 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.234631 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.236924 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f3052a777d2db499f91f40a7517c0f2cd48191aaf6ebc8b17f1c114c2623fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41e994b69b7552394d02f9225cf931743737e1bfcd0448aff5f860ab27afa00b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.250964 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.263066 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.277935 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.299938 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:06Z\\\",\\\"message\\\":\\\"mplate:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.153\\\\\\\", Port:5443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1210 18:56:06.639052 6459 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z]\\\\nI1210 18:56:06.638925 6459 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-scheduler-operator for network=default : 3.861754ms\\\\nI1210 18:56:06.639267 6459 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.311534 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-74lmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"472ff1d2-0936-4ee0-b738-dc26e2c03559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-74lmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.324225 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.337264 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.337303 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.337315 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.337337 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.337350 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.337377 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.350962 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:08Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.439181 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.439226 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.439237 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.439253 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.439262 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.541307 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.541378 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.541402 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.541431 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.541451 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.644710 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.644745 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.644753 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.644768 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.644776 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.747430 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.747478 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.747495 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.747517 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.747533 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.788038 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.788142 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:08 crc kubenswrapper[4828]: E1210 18:56:08.788222 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.788244 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.788301 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:08 crc kubenswrapper[4828]: E1210 18:56:08.788484 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:08 crc kubenswrapper[4828]: E1210 18:56:08.788712 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:08 crc kubenswrapper[4828]: E1210 18:56:08.788856 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.849935 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.849987 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.850004 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.850026 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.850043 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.852314 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs\") pod \"network-metrics-daemon-74lmx\" (UID: \"472ff1d2-0936-4ee0-b738-dc26e2c03559\") " pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:08 crc kubenswrapper[4828]: E1210 18:56:08.852449 4828 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:08 crc kubenswrapper[4828]: E1210 18:56:08.852490 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs podName:472ff1d2-0936-4ee0-b738-dc26e2c03559 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:24.852477084 +0000 UTC m=+65.363088089 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs") pod "network-metrics-daemon-74lmx" (UID: "472ff1d2-0936-4ee0-b738-dc26e2c03559") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.952338 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.952367 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.952378 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.952390 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:08 crc kubenswrapper[4828]: I1210 18:56:08.952399 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:08Z","lastTransitionTime":"2025-12-10T18:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.055171 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.055226 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.055240 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.055262 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.055292 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.157590 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.157645 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.157723 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.157742 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.157754 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.260256 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.260331 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.260349 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.260376 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.260393 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.364242 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.364341 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.364360 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.364382 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.364400 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.467763 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.467856 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.467874 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.467894 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.467935 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.570128 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.570199 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.570209 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.570223 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.570233 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.672784 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.672884 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.672905 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.672934 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.672954 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.777070 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.777130 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.777150 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.777178 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.777195 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.810976 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:09Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.841037 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:06Z\\\",\\\"message\\\":\\\"mplate:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.153\\\\\\\", Port:5443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1210 18:56:06.639052 6459 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z]\\\\nI1210 18:56:06.638925 6459 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-scheduler-operator for network=default : 3.861754ms\\\\nI1210 18:56:06.639267 6459 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:09Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.867433 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-74lmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"472ff1d2-0936-4ee0-b738-dc26e2c03559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-74lmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:09Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.879601 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.879666 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.879684 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.879715 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.879737 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.889673 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:09Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.909876 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:09Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.925461 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:09Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.946224 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:09Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.966112 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:09Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.982578 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.982627 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.982642 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.982666 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.982683 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:09Z","lastTransitionTime":"2025-12-10T18:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:09 crc kubenswrapper[4828]: I1210 18:56:09.988829 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:09Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.005031 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:10Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.029389 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:10Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.051619 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:10Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.067852 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:10Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.083657 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f3052a777d2db499f91f40a7517c0f2cd48191aaf6ebc8b17f1c114c2623fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41e994b69b7552394d02f9225cf931743737e1bfcd0448aff5f860ab27afa00b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:10Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.085642 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.085750 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.085786 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.085868 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.085903 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.104330 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:10Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.122558 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:10Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.189983 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.190036 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.190052 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.190076 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.190097 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.294270 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.294598 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.294642 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.294679 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.294709 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.398294 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.398437 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.398455 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.398481 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.398499 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.501530 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.501641 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.501666 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.501699 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.501719 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.605415 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.605504 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.605523 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.605556 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.605579 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.673959 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.674171 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:10 crc kubenswrapper[4828]: E1210 18:56:10.674228 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:56:42.67418527 +0000 UTC m=+83.184796275 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.674463 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:10 crc kubenswrapper[4828]: E1210 18:56:10.674680 4828 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:56:10 crc kubenswrapper[4828]: E1210 18:56:10.674890 4828 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:56:10 crc kubenswrapper[4828]: E1210 18:56:10.674945 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:42.674896713 +0000 UTC m=+83.185507908 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:56:10 crc kubenswrapper[4828]: E1210 18:56:10.674997 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:42.674976545 +0000 UTC m=+83.185587770 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.709244 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.709332 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.709356 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.709392 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.709418 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.775628 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.775678 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:10 crc kubenswrapper[4828]: E1210 18:56:10.775780 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:56:10 crc kubenswrapper[4828]: E1210 18:56:10.775827 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:56:10 crc kubenswrapper[4828]: E1210 18:56:10.775838 4828 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:56:10 crc kubenswrapper[4828]: E1210 18:56:10.775841 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:56:10 crc kubenswrapper[4828]: E1210 18:56:10.775866 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:56:10 crc kubenswrapper[4828]: E1210 18:56:10.775880 4828 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:56:10 crc kubenswrapper[4828]: E1210 18:56:10.775893 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:42.775878102 +0000 UTC m=+83.286489107 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:56:10 crc kubenswrapper[4828]: E1210 18:56:10.775918 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:42.775905923 +0000 UTC m=+83.286516948 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.788701 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.788777 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.788728 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.788717 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:10 crc kubenswrapper[4828]: E1210 18:56:10.789031 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:10 crc kubenswrapper[4828]: E1210 18:56:10.789348 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:10 crc kubenswrapper[4828]: E1210 18:56:10.789490 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:10 crc kubenswrapper[4828]: E1210 18:56:10.789642 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.812506 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.812587 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.812614 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.812650 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.812675 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.915439 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.915475 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.915484 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.915497 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:10 crc kubenswrapper[4828]: I1210 18:56:10.915507 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:10Z","lastTransitionTime":"2025-12-10T18:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.018913 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.018970 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.018982 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.019002 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.019014 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.121832 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.121918 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.121941 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.121999 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.122025 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.224992 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.225041 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.225055 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.225079 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.225096 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.327194 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.327251 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.327268 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.327290 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.327305 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.430844 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.431287 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.431309 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.431336 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.431354 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.534602 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.534667 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.534681 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.534704 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.534723 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.637164 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.637200 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.637209 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.637222 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.637231 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.729984 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.739687 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.739731 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.739741 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.739757 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.739771 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.742962 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.745379 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:11Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.768649 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:11Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.786766 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:11Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.807955 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:11Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.827555 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:11Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.843377 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.843460 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.843479 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.843507 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.843524 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.848957 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:11Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.867773 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:11Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.885727 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:11Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.902771 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f3052a777d2db499f91f40a7517c0f2cd48191aaf6ebc8b17f1c114c2623fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41e994b69b7552394d02f9225cf931743737e1bfcd0448aff5f860ab27afa00b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:11Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.917470 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:11Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.934407 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:11Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.946389 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.946447 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.946461 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.946477 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.946488 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:11Z","lastTransitionTime":"2025-12-10T18:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.955562 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:11Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.979278 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:06Z\\\",\\\"message\\\":\\\"mplate:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.153\\\\\\\", Port:5443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1210 18:56:06.639052 6459 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z]\\\\nI1210 18:56:06.638925 6459 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-scheduler-operator for network=default : 3.861754ms\\\\nI1210 18:56:06.639267 6459 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:11Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:11 crc kubenswrapper[4828]: I1210 18:56:11.991472 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-74lmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"472ff1d2-0936-4ee0-b738-dc26e2c03559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-74lmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:11Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.010293 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:12Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.027155 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:12Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.049162 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.049228 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.049242 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.049258 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.049267 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.152311 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.152361 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.152373 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.152429 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.152443 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.254928 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.255008 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.255027 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.255057 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.255079 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.358912 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.358989 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.359007 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.359039 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.359061 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.422353 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.422445 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.422466 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.422496 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.422519 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4828]: E1210 18:56:12.440639 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:12Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.445361 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.445421 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.445440 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.445470 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.445493 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4828]: E1210 18:56:12.464112 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:12Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.469701 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.469757 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.469776 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.469820 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.469836 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4828]: E1210 18:56:12.488471 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:12Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.496013 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.496130 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.496157 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.496192 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.496216 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4828]: E1210 18:56:12.519218 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:12Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.524614 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.524681 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.524700 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.524726 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.524745 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4828]: E1210 18:56:12.538543 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:12Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:12 crc kubenswrapper[4828]: E1210 18:56:12.538683 4828 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.540359 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.540406 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.540450 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.540470 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.540481 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.642599 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.642633 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.642645 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.642660 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.642668 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.750998 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.751073 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.751093 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.751121 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.751153 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.788886 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.788961 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.789047 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.789107 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:12 crc kubenswrapper[4828]: E1210 18:56:12.789181 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:12 crc kubenswrapper[4828]: E1210 18:56:12.789309 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:12 crc kubenswrapper[4828]: E1210 18:56:12.789398 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:12 crc kubenswrapper[4828]: E1210 18:56:12.789541 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.854166 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.854202 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.854213 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.854227 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.854235 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.957155 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.957230 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.957255 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.957285 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:12 crc kubenswrapper[4828]: I1210 18:56:12.957309 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:12Z","lastTransitionTime":"2025-12-10T18:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.060964 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.061013 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.061027 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.061049 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.061065 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.163493 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.163536 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.163547 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.163564 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.163575 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.265863 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.265949 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.265975 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.266006 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.266030 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.368899 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.369000 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.369021 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.369085 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.369106 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.471303 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.471344 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.471356 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.471372 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.471386 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.574647 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.575086 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.575281 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.575483 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.575612 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.678220 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.678276 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.678285 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.678299 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.678310 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.781370 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.781410 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.781419 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.781436 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.781445 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.883395 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.883456 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.883475 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.883499 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.883518 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.986301 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.986347 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.986355 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.986369 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:13 crc kubenswrapper[4828]: I1210 18:56:13.986378 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:13Z","lastTransitionTime":"2025-12-10T18:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.089334 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.089383 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.089395 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.089414 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.089426 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.191915 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.191957 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.191966 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.191979 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.191991 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.295201 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.295284 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.295306 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.295334 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.295355 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.398665 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.398735 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.398766 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.398841 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.398865 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.501968 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.502041 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.502055 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.502079 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.502095 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.605413 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.605457 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.605472 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.605497 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.605509 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.708045 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.708092 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.708108 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.708129 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.708146 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.787726 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.787782 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.787747 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.787971 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:14 crc kubenswrapper[4828]: E1210 18:56:14.787966 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:14 crc kubenswrapper[4828]: E1210 18:56:14.788080 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:14 crc kubenswrapper[4828]: E1210 18:56:14.788286 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:14 crc kubenswrapper[4828]: E1210 18:56:14.788369 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.809974 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.810031 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.810042 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.810056 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.810065 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.912730 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.912770 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.912780 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.912816 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:14 crc kubenswrapper[4828]: I1210 18:56:14.912828 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:14Z","lastTransitionTime":"2025-12-10T18:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.015197 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.015240 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.015249 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.015264 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.015274 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.118881 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.118970 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.118995 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.119025 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.119049 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.222937 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.223021 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.223041 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.223068 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.223088 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.326101 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.326187 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.326213 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.326246 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.326268 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.429212 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.429264 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.429276 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.429295 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.429307 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.532664 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.532745 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.532764 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.532788 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.532834 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.636461 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.636520 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.636537 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.636565 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.636583 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.739216 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.739298 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.739317 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.739349 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.739371 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.842767 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.842850 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.842863 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.842882 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.842916 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.946697 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.946739 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.946751 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.946765 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:15 crc kubenswrapper[4828]: I1210 18:56:15.946774 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:15Z","lastTransitionTime":"2025-12-10T18:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.050730 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.050873 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.050900 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.050929 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.050949 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.153848 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.153952 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.153972 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.154018 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.154042 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.257792 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.257895 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.257912 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.257941 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.257959 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.360889 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.360944 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.360960 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.360984 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.361006 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.464456 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.464507 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.464524 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.464546 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.464562 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.567647 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.567709 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.567744 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.567781 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.567844 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.671191 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.671238 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.671250 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.671265 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.671277 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.773788 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.773882 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.773896 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.773924 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.773937 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.787925 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.787992 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.787995 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.787922 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:16 crc kubenswrapper[4828]: E1210 18:56:16.788129 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:16 crc kubenswrapper[4828]: E1210 18:56:16.788277 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:16 crc kubenswrapper[4828]: E1210 18:56:16.788403 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:16 crc kubenswrapper[4828]: E1210 18:56:16.788567 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.876669 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.876737 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.876754 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.876782 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.876826 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.979214 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.979281 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.979295 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.979314 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:16 crc kubenswrapper[4828]: I1210 18:56:16.979325 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:16Z","lastTransitionTime":"2025-12-10T18:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.082549 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.082621 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.082644 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.082670 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.082688 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.186082 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.186161 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.186180 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.186211 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.186261 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.289312 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.289543 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.289567 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.289591 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.289613 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.393240 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.393323 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.393342 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.393380 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.393407 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.496546 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.496688 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.496732 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.496772 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.496842 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.599549 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.599597 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.599613 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.599638 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.599655 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.702908 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.702976 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.702993 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.703016 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.703033 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.805326 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.805390 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.805407 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.805430 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.805447 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.907929 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.907967 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.907974 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.907988 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:17 crc kubenswrapper[4828]: I1210 18:56:17.907997 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:17Z","lastTransitionTime":"2025-12-10T18:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.010038 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.010115 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.010138 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.010156 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.010165 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.113030 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.113078 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.113091 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.113109 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.113125 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.217124 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.217186 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.217196 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.217212 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.217222 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.320530 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.320574 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.320582 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.320597 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.320606 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.423177 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.423222 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.423233 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.423251 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.423261 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.525885 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.525934 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.525946 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.525964 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.525976 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.628614 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.628667 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.628679 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.628697 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.628712 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.731313 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.731368 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.731389 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.731414 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.731433 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.787737 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.787785 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.787743 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.787909 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:18 crc kubenswrapper[4828]: E1210 18:56:18.787968 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:18 crc kubenswrapper[4828]: E1210 18:56:18.788022 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:18 crc kubenswrapper[4828]: E1210 18:56:18.788254 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:18 crc kubenswrapper[4828]: E1210 18:56:18.788404 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.833784 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.833846 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.833857 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.833874 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.833889 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.937057 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.937122 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.937140 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.937163 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:18 crc kubenswrapper[4828]: I1210 18:56:18.937181 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:18Z","lastTransitionTime":"2025-12-10T18:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.040533 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.040591 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.040608 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.040630 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.040649 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.143181 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.143222 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.143232 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.143248 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.143260 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.246403 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.246471 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.246496 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.246527 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.246549 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.350176 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.350260 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.350279 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.350303 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.350321 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.453318 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.453362 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.453371 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.453385 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.453394 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.555922 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.555994 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.556018 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.556047 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.556068 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.658140 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.658177 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.658187 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.658204 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.658216 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.760760 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.760829 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.760838 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.760873 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.760883 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.798502 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.811867 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.822714 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.833665 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f3052a777d2db499f91f40a7517c0f2cd48191aaf6ebc8b17f1c114c2623fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41e994b69b7552394d02f9225cf931743737e1bfcd0448aff5f860ab27afa00b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.845757 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.862413 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.862451 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.862463 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.862506 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.862520 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.866182 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.883942 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.912793 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:06Z\\\",\\\"message\\\":\\\"mplate:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.153\\\\\\\", Port:5443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1210 18:56:06.639052 6459 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z]\\\\nI1210 18:56:06.638925 6459 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-scheduler-operator for network=default : 3.861754ms\\\\nI1210 18:56:06.639267 6459 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.927948 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-74lmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"472ff1d2-0936-4ee0-b738-dc26e2c03559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-74lmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.947015 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.961089 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.965047 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.965070 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.965081 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.965098 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.965117 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:19Z","lastTransitionTime":"2025-12-10T18:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.976580 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:19 crc kubenswrapper[4828]: I1210 18:56:19.992916 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:19Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.005995 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:20Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.019463 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:20Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.029647 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:20Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.038905 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60cfa1f-f033-4ea6-8785-b00708e0af0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1dcfb31658e3faf516dbebd1e00d1f8dc701e5c6b73831a52648effb712db12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66fc64d5883f4b746a0636317bd472362758bdbfa1c5ab6a0a327692256e064a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b044808efab4ceca84bf0256a92f6559ea57b15a0d2a6c83756f489accbefdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:20Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.068048 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.068082 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.068092 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.068106 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.068114 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.170332 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.170395 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.170412 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.170435 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.170451 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.272790 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.272918 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.272943 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.272975 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.272998 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.376015 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.376073 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.376094 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.376125 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.376148 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.478457 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.478489 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.478499 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.478514 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.478524 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.581483 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.581519 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.581530 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.581546 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.581559 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.683854 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.683926 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.683948 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.683982 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.684004 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.786884 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.786925 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.786934 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.786946 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.786955 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.787840 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.787896 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.787951 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:20 crc kubenswrapper[4828]: E1210 18:56:20.788006 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.787864 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:20 crc kubenswrapper[4828]: E1210 18:56:20.788236 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:20 crc kubenswrapper[4828]: E1210 18:56:20.788281 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:20 crc kubenswrapper[4828]: E1210 18:56:20.788403 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.889367 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.889421 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.889460 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.889485 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.889502 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.991722 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.991828 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.991847 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.991870 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:20 crc kubenswrapper[4828]: I1210 18:56:20.991887 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:20Z","lastTransitionTime":"2025-12-10T18:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.094705 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.094789 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.094833 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.094852 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.094867 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.199277 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.199328 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.199340 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.199360 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.199372 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.302499 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.302562 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.302581 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.302608 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.302624 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.405645 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.405708 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.405726 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.405756 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.405777 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.508823 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.508872 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.508884 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.508902 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.508916 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.611009 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.611052 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.611064 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.611080 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.611092 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.713292 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.713555 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.713667 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.713778 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.713915 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.815835 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.815869 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.815881 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.815898 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.815912 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.919077 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.919132 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.919144 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.919162 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:21 crc kubenswrapper[4828]: I1210 18:56:21.919175 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:21Z","lastTransitionTime":"2025-12-10T18:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.021814 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.021853 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.021862 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.021875 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.021884 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.124914 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.124956 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.124969 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.124989 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.125004 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.227121 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.227403 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.227480 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.227557 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.227622 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.329747 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.330051 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.330145 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.330287 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.330387 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.433567 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.433606 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.433614 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.433628 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.433639 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.535491 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.535522 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.535530 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.535600 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.535609 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.579085 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.579126 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.579137 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.579154 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.579169 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4828]: E1210 18:56:22.591076 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:22Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.598731 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.598778 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.598788 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.598814 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.598824 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4828]: E1210 18:56:22.611175 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:22Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.614914 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.614972 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.614983 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.615000 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.615010 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4828]: E1210 18:56:22.631164 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:22Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.634329 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.634374 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.634386 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.634405 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.634416 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4828]: E1210 18:56:22.644879 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:22Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.648360 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.648386 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.648396 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.648407 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.648415 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4828]: E1210 18:56:22.660435 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:22Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:22 crc kubenswrapper[4828]: E1210 18:56:22.660546 4828 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.662276 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.662304 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.662314 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.662328 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.662336 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.764874 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.764944 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.764959 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.764977 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.764988 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.788091 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.788128 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.788189 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:22 crc kubenswrapper[4828]: E1210 18:56:22.788224 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:22 crc kubenswrapper[4828]: E1210 18:56:22.788373 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:22 crc kubenswrapper[4828]: E1210 18:56:22.788417 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.788564 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:22 crc kubenswrapper[4828]: E1210 18:56:22.788729 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.798255 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.867199 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.867243 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.867255 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.867271 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.867283 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.970241 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.970284 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.970295 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.970311 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:22 crc kubenswrapper[4828]: I1210 18:56:22.970323 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:22Z","lastTransitionTime":"2025-12-10T18:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.072657 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.072717 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.072730 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.072756 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.072768 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.174642 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.174676 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.174687 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.174701 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.174711 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.277029 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.277071 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.277083 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.277098 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.277108 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.378945 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.378971 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.378979 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.378991 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.379000 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.481252 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.481288 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.481298 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.481315 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.481327 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.583501 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.583544 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.583555 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.583571 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.583582 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.685445 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.685503 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.685523 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.685546 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.685563 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.788084 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.788375 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.788448 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.788516 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.788579 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.788598 4828 scope.go:117] "RemoveContainer" containerID="fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00" Dec 10 18:56:23 crc kubenswrapper[4828]: E1210 18:56:23.789124 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.899244 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.899304 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.899317 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.899334 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:23 crc kubenswrapper[4828]: I1210 18:56:23.899668 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:23Z","lastTransitionTime":"2025-12-10T18:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.002677 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.002715 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.002729 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.002748 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.002762 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.104671 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.104703 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.104714 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.104729 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.104739 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.206763 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.206834 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.206846 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.206862 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.206873 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.309591 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.310165 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.310340 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.310518 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.310674 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.413857 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.413910 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.413923 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.413943 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.413957 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.516351 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.516406 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.516423 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.516448 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.516465 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.618519 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.618548 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.618556 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.618569 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.618620 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.720888 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.720928 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.721157 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.721177 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.721221 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.788545 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.788631 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:24 crc kubenswrapper[4828]: E1210 18:56:24.788678 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.788854 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:24 crc kubenswrapper[4828]: E1210 18:56:24.788927 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.788967 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:24 crc kubenswrapper[4828]: E1210 18:56:24.789161 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:24 crc kubenswrapper[4828]: E1210 18:56:24.789228 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.823470 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.823499 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.823506 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.823522 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.823531 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.917703 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs\") pod \"network-metrics-daemon-74lmx\" (UID: \"472ff1d2-0936-4ee0-b738-dc26e2c03559\") " pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:24 crc kubenswrapper[4828]: E1210 18:56:24.917870 4828 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:24 crc kubenswrapper[4828]: E1210 18:56:24.917930 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs podName:472ff1d2-0936-4ee0-b738-dc26e2c03559 nodeName:}" failed. No retries permitted until 2025-12-10 18:56:56.917912552 +0000 UTC m=+97.428523557 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs") pod "network-metrics-daemon-74lmx" (UID: "472ff1d2-0936-4ee0-b738-dc26e2c03559") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.926502 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.926574 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.926588 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.926609 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:24 crc kubenswrapper[4828]: I1210 18:56:24.926622 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:24Z","lastTransitionTime":"2025-12-10T18:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.028667 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.028703 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.028713 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.028728 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.028739 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.130680 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.130751 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.130770 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.130829 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.130850 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.232822 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.232861 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.232874 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.232890 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.232902 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.335340 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.335406 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.335425 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.335450 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.335468 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.438110 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.438171 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.438192 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.438226 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.438248 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.540666 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.540716 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.540725 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.540741 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.540750 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.643297 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.643335 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.643343 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.643357 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.643368 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.744855 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.744889 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.744902 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.744919 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.744933 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.849024 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.849118 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.849136 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.849160 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.849177 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.951869 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.951912 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.951923 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.951940 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:25 crc kubenswrapper[4828]: I1210 18:56:25.951954 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:25Z","lastTransitionTime":"2025-12-10T18:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.054513 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.054589 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.054599 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.054613 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.054621 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.157659 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.157712 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.157724 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.157744 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.157757 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.175343 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xdxgd_c6e55203-65ae-4767-9a96-f309bc41ddf3/kube-multus/0.log" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.175422 4828 generic.go:334] "Generic (PLEG): container finished" podID="c6e55203-65ae-4767-9a96-f309bc41ddf3" containerID="6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085" exitCode=1 Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.175472 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xdxgd" event={"ID":"c6e55203-65ae-4767-9a96-f309bc41ddf3","Type":"ContainerDied","Data":"6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085"} Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.176181 4828 scope.go:117] "RemoveContainer" containerID="6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.210843 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:06Z\\\",\\\"message\\\":\\\"mplate:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.153\\\\\\\", Port:5443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1210 18:56:06.639052 6459 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z]\\\\nI1210 18:56:06.638925 6459 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-scheduler-operator for network=default : 3.861754ms\\\\nI1210 18:56:06.639267 6459 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.221819 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-74lmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"472ff1d2-0936-4ee0-b738-dc26e2c03559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-74lmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.234981 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.250774 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.260606 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.260634 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.260643 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.260673 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.260682 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.264173 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.277326 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1c81c13-3c65-4b70-b50a-177b5d07c0a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3f41b839ee9d8b2b5083bd8b5ba463e892011fcfb9052449f9198c6bba5d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://215812f5a8013b31678b85a9fa5232c2da907e04f587be31f9508ea39182976b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://215812f5a8013b31678b85a9fa5232c2da907e04f587be31f9508ea39182976b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.288449 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.298954 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.310964 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.322140 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.331455 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60cfa1f-f033-4ea6-8785-b00708e0af0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1dcfb31658e3faf516dbebd1e00d1f8dc701e5c6b73831a52648effb712db12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66fc64d5883f4b746a0636317bd472362758bdbfa1c5ab6a0a327692256e064a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b044808efab4ceca84bf0256a92f6559ea57b15a0d2a6c83756f489accbefdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.342949 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.352943 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:26Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:25Z\\\",\\\"message\\\":\\\"2025-12-10T18:55:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c10c5c27-b43d-4d94-9259-eaa0eddec29a\\\\n2025-12-10T18:55:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c10c5c27-b43d-4d94-9259-eaa0eddec29a to /host/opt/cni/bin/\\\\n2025-12-10T18:55:40Z [verbose] multus-daemon started\\\\n2025-12-10T18:55:40Z [verbose] Readiness Indicator file check\\\\n2025-12-10T18:56:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.361678 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.364534 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.364558 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.364566 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.364580 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.364588 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.371704 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f3052a777d2db499f91f40a7517c0f2cd48191aaf6ebc8b17f1c114c2623fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41e994b69b7552394d02f9225cf931743737e1bfcd0448aff5f860ab27afa00b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.381548 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.389959 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.398961 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:26Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.467038 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.467074 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.467085 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.467103 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.467114 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.569326 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.569357 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.569399 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.569413 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.569424 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.671166 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.671393 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.671460 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.671531 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.671587 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.774493 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.774531 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.774543 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.774560 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.774572 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.787685 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.787730 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.787688 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.787686 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:26 crc kubenswrapper[4828]: E1210 18:56:26.787790 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:26 crc kubenswrapper[4828]: E1210 18:56:26.787897 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:26 crc kubenswrapper[4828]: E1210 18:56:26.787998 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:26 crc kubenswrapper[4828]: E1210 18:56:26.788094 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.876431 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.876467 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.876475 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.876489 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.876499 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.978383 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.978431 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.978439 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.978450 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:26 crc kubenswrapper[4828]: I1210 18:56:26.978459 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:26Z","lastTransitionTime":"2025-12-10T18:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.080502 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.080535 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.080544 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.080557 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.080566 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.179689 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xdxgd_c6e55203-65ae-4767-9a96-f309bc41ddf3/kube-multus/0.log" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.179742 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xdxgd" event={"ID":"c6e55203-65ae-4767-9a96-f309bc41ddf3","Type":"ContainerStarted","Data":"bbe415208b7c4a043d3a9466d81931a16b2d600a3b5605eb89c7d2c2ffb1b71b"} Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.183073 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.183111 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.183120 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.183136 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.183145 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.191032 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.202265 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60cfa1f-f033-4ea6-8785-b00708e0af0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1dcfb31658e3faf516dbebd1e00d1f8dc701e5c6b73831a52648effb712db12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66fc64d5883f4b746a0636317bd472362758bdbfa1c5ab6a0a327692256e064a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b044808efab4ceca84bf0256a92f6559ea57b15a0d2a6c83756f489accbefdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.213074 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.223339 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.230943 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.239259 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f3052a777d2db499f91f40a7517c0f2cd48191aaf6ebc8b17f1c114c2623fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41e994b69b7552394d02f9225cf931743737e1bfcd0448aff5f860ab27afa00b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.249323 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.260078 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.270932 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.281600 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe415208b7c4a043d3a9466d81931a16b2d600a3b5605eb89c7d2c2ffb1b71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:25Z\\\",\\\"message\\\":\\\"2025-12-10T18:55:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c10c5c27-b43d-4d94-9259-eaa0eddec29a\\\\n2025-12-10T18:55:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c10c5c27-b43d-4d94-9259-eaa0eddec29a to /host/opt/cni/bin/\\\\n2025-12-10T18:55:40Z [verbose] multus-daemon started\\\\n2025-12-10T18:55:40Z [verbose] Readiness Indicator file check\\\\n2025-12-10T18:56:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.285018 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.285053 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.285062 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.285077 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.285088 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.294240 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-74lmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"472ff1d2-0936-4ee0-b738-dc26e2c03559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-74lmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.308322 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.320598 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.336238 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.377296 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:06Z\\\",\\\"message\\\":\\\"mplate:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.153\\\\\\\", Port:5443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1210 18:56:06.639052 6459 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z]\\\\nI1210 18:56:06.638925 6459 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-scheduler-operator for network=default : 3.861754ms\\\\nI1210 18:56:06.639267 6459 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.387021 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.387049 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.387057 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.387073 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.387081 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.394062 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1c81c13-3c65-4b70-b50a-177b5d07c0a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3f41b839ee9d8b2b5083bd8b5ba463e892011fcfb9052449f9198c6bba5d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://215812f5a8013b31678b85a9fa5232c2da907e04f587be31f9508ea39182976b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://215812f5a8013b31678b85a9fa5232c2da907e04f587be31f9508ea39182976b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.404963 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.416193 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:27Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.489140 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.489179 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.489190 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.489206 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.489218 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.591582 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.591621 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.591629 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.591642 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.591651 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.693473 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.693501 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.693512 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.693527 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.693538 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.794994 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.795033 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.795046 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.795059 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.795070 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.898006 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.898046 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.898055 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.898071 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:27 crc kubenswrapper[4828]: I1210 18:56:27.898081 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:27Z","lastTransitionTime":"2025-12-10T18:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.000616 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.000672 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.000683 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.000720 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.000732 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.102867 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.102898 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.102905 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.102917 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.102926 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.204635 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.204670 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.204678 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.204692 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.204702 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.308192 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.308224 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.308235 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.308249 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.308260 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.410422 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.410452 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.410460 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.410473 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.410483 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.512560 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.512600 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.512609 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.512625 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.512633 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.614462 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.614501 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.614510 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.614525 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.614538 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.716903 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.716976 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.717006 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.717041 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.717064 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.787957 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.788055 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.788070 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:28 crc kubenswrapper[4828]: E1210 18:56:28.788163 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.788244 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:28 crc kubenswrapper[4828]: E1210 18:56:28.788289 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:28 crc kubenswrapper[4828]: E1210 18:56:28.788390 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:28 crc kubenswrapper[4828]: E1210 18:56:28.788427 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.820152 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.820187 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.820197 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.820214 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.820225 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.923462 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.923507 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.923519 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.923537 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:28 crc kubenswrapper[4828]: I1210 18:56:28.923551 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:28Z","lastTransitionTime":"2025-12-10T18:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.026488 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.026528 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.026538 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.026552 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.026563 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.128982 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.129040 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.129054 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.129073 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.129085 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.231354 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.231401 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.231412 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.231427 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.231436 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.333360 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.333409 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.333417 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.333432 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.333442 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.435597 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.435648 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.435663 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.435683 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.435695 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.538104 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.538202 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.538220 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.538243 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.538261 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.640542 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.640579 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.640587 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.640604 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.640612 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.743521 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.743571 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.743582 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.743599 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.743610 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.798022 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.813462 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60cfa1f-f033-4ea6-8785-b00708e0af0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1dcfb31658e3faf516dbebd1e00d1f8dc701e5c6b73831a52648effb712db12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66fc64d5883f4b746a0636317bd472362758bdbfa1c5ab6a0a327692256e064a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b044808efab4ceca84bf0256a92f6559ea57b15a0d2a6c83756f489accbefdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.826274 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.839773 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.846365 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.846405 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.846416 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.846433 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.846446 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.855172 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.866914 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f3052a777d2db499f91f40a7517c0f2cd48191aaf6ebc8b17f1c114c2623fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41e994b69b7552394d02f9225cf931743737e1bfcd0448aff5f860ab27afa00b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.880684 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.891754 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.903500 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.916320 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe415208b7c4a043d3a9466d81931a16b2d600a3b5605eb89c7d2c2ffb1b71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:25Z\\\",\\\"message\\\":\\\"2025-12-10T18:55:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c10c5c27-b43d-4d94-9259-eaa0eddec29a\\\\n2025-12-10T18:55:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c10c5c27-b43d-4d94-9259-eaa0eddec29a to /host/opt/cni/bin/\\\\n2025-12-10T18:55:40Z [verbose] multus-daemon started\\\\n2025-12-10T18:55:40Z [verbose] Readiness Indicator file check\\\\n2025-12-10T18:56:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.925312 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-74lmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"472ff1d2-0936-4ee0-b738-dc26e2c03559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-74lmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.938357 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.949839 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.949882 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.949894 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.949910 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.949926 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:29Z","lastTransitionTime":"2025-12-10T18:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.950292 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.964679 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.983159 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:06Z\\\",\\\"message\\\":\\\"mplate:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.153\\\\\\\", Port:5443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1210 18:56:06.639052 6459 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z]\\\\nI1210 18:56:06.638925 6459 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-scheduler-operator for network=default : 3.861754ms\\\\nI1210 18:56:06.639267 6459 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:29 crc kubenswrapper[4828]: I1210 18:56:29.992557 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1c81c13-3c65-4b70-b50a-177b5d07c0a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3f41b839ee9d8b2b5083bd8b5ba463e892011fcfb9052449f9198c6bba5d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://215812f5a8013b31678b85a9fa5232c2da907e04f587be31f9508ea39182976b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://215812f5a8013b31678b85a9fa5232c2da907e04f587be31f9508ea39182976b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:29Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.003638 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:30Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.016008 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:30Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.051734 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.051767 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.051778 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.051792 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.051819 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.153560 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.153596 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.153605 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.153618 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.153627 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.255741 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.255781 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.255791 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.255826 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.255840 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.358344 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.358380 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.358389 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.358403 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.358411 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.460139 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.460179 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.460189 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.460202 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.460210 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.562963 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.563035 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.563052 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.563081 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.563100 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.665030 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.665073 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.665087 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.665106 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.665119 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.767576 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.767632 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.767651 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.767673 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.767692 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.788228 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.788321 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:30 crc kubenswrapper[4828]: E1210 18:56:30.788360 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.788235 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.788238 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:30 crc kubenswrapper[4828]: E1210 18:56:30.788492 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:30 crc kubenswrapper[4828]: E1210 18:56:30.788650 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:30 crc kubenswrapper[4828]: E1210 18:56:30.788694 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.871097 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.871154 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.871171 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.871227 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.871248 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.973120 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.973156 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.973166 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.973182 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:30 crc kubenswrapper[4828]: I1210 18:56:30.973192 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:30Z","lastTransitionTime":"2025-12-10T18:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.075485 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.075518 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.075530 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.075545 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.075556 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.178269 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.178309 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.178321 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.178337 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.178349 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.280956 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.280993 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.281005 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.281021 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.281034 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.383264 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.383298 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.383311 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.383329 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.383341 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.485325 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.485370 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.485384 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.485405 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.485420 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.588196 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.588245 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.588259 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.588275 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.588288 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.690084 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.690116 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.690123 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.690137 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.690145 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.791729 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.791770 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.791780 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.791792 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.791820 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.894268 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.894305 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.894317 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.894332 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.894342 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.996528 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.996579 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.996587 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.996603 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:31 crc kubenswrapper[4828]: I1210 18:56:31.996614 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:31Z","lastTransitionTime":"2025-12-10T18:56:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.099250 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.099294 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.099306 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.099325 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.099335 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.201393 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.201435 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.201446 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.201462 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.201475 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.304622 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.304673 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.304688 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.304720 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.304736 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.408180 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.408266 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.408279 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.408295 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.408307 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.510653 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.510694 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.510703 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.510717 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.510728 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.612442 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.612473 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.612481 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.612494 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.612502 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.714772 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.714840 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.714855 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.714872 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.714883 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.788103 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.788156 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.788156 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.788172 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:32 crc kubenswrapper[4828]: E1210 18:56:32.788271 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:32 crc kubenswrapper[4828]: E1210 18:56:32.788375 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:32 crc kubenswrapper[4828]: E1210 18:56:32.788515 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:32 crc kubenswrapper[4828]: E1210 18:56:32.788609 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.816748 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.816823 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.816834 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.816858 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.816870 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.919052 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.919088 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.919097 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.919113 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:32 crc kubenswrapper[4828]: I1210 18:56:32.919124 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:32Z","lastTransitionTime":"2025-12-10T18:56:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.021317 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.021353 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.021364 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.021381 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.021392 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.023212 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.023240 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.023250 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.023260 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.023268 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4828]: E1210 18:56:33.034646 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:33Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.037599 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.037640 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.037649 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.037668 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.037678 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4828]: E1210 18:56:33.048656 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:33Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.051924 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.051965 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.051999 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.052017 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.052028 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4828]: E1210 18:56:33.066920 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:33Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.070596 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.070632 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.070640 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.070652 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.070661 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4828]: E1210 18:56:33.083384 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:33Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.086354 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.086404 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.086416 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.086435 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.086448 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4828]: E1210 18:56:33.097464 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:33Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:33Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:33 crc kubenswrapper[4828]: E1210 18:56:33.097580 4828 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.123052 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.123091 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.123099 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.123111 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.123119 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.225471 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.225531 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.225551 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.225576 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.225593 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.328552 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.328600 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.328613 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.328630 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.328643 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.431441 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.431504 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.431520 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.431544 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.431563 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.534218 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.534283 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.534301 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.534325 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.534369 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.637557 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.637622 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.637639 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.637664 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.637686 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.741015 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.741087 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.741104 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.741133 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.741151 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.844444 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.844484 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.844495 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.844512 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.844524 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.947627 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.947670 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.947686 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.947709 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:33 crc kubenswrapper[4828]: I1210 18:56:33.947725 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:33Z","lastTransitionTime":"2025-12-10T18:56:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.050091 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.050120 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.050127 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.050139 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.050147 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.153450 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.153500 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.153516 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.153535 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.153550 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.256078 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.256152 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.256168 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.256190 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.256204 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.359573 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.359634 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.359650 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.359673 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.359691 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.462696 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.462764 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.462787 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.462848 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.462871 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.566060 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.566119 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.566134 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.566157 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.566175 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.669632 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.669710 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.669743 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.669771 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.669792 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.772982 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.773059 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.773083 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.773109 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.773127 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.788608 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.788661 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.788708 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.788623 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:34 crc kubenswrapper[4828]: E1210 18:56:34.788860 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:34 crc kubenswrapper[4828]: E1210 18:56:34.789045 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:34 crc kubenswrapper[4828]: E1210 18:56:34.789189 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:34 crc kubenswrapper[4828]: E1210 18:56:34.789351 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.876175 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.876315 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.876336 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.876364 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.876382 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.980136 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.980211 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.980228 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.980268 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:34 crc kubenswrapper[4828]: I1210 18:56:34.980286 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:34Z","lastTransitionTime":"2025-12-10T18:56:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.082058 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.082102 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.082112 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.082129 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.082141 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.184222 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.184283 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.184301 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.184326 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.184343 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.286525 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.286592 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.286609 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.286633 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.286649 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.389722 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.389834 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.389852 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.389876 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.389893 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.492906 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.492971 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.492988 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.493014 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.493032 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.595436 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.595482 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.595495 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.595512 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.595525 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.698019 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.698058 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.698068 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.698082 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.698091 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.801000 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.801089 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.801110 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.801136 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.801155 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.904344 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.904419 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.904436 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.904459 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:35 crc kubenswrapper[4828]: I1210 18:56:35.904477 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:35Z","lastTransitionTime":"2025-12-10T18:56:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.006975 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.007055 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.007078 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.007108 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.007136 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.109779 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.109842 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.109853 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.109869 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.109877 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.212636 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.212680 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.212689 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.212708 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.212717 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.315267 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.315304 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.315312 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.315326 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.315336 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.418185 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.418252 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.418269 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.418301 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.418320 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.521518 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.521619 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.521637 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.521663 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.521684 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.624937 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.625004 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.625026 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.625055 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.625074 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.728269 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.728335 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.728351 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.728377 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.728394 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.788596 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.788647 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.788617 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:36 crc kubenswrapper[4828]: E1210 18:56:36.788874 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.789011 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:36 crc kubenswrapper[4828]: E1210 18:56:36.789316 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:36 crc kubenswrapper[4828]: E1210 18:56:36.789397 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:36 crc kubenswrapper[4828]: E1210 18:56:36.789501 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.832029 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.832092 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.832108 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.832134 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.832153 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.935383 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.935454 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.935473 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.935502 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:36 crc kubenswrapper[4828]: I1210 18:56:36.935522 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:36Z","lastTransitionTime":"2025-12-10T18:56:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.041982 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.042056 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.042068 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.042112 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.042123 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.144217 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.144287 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.144302 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.144325 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.144342 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.247978 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.248028 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.248037 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.248059 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.248072 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.351343 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.351399 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.351415 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.351435 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.351449 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.454163 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.454225 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.454236 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.454254 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.454265 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.556758 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.556822 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.556869 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.556888 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.556899 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.659280 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.659362 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.659387 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.659421 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.659447 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.762123 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.762193 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.762210 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.762237 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.762259 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.865012 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.865049 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.865061 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.865077 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.865087 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.967355 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.967417 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.967431 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.967448 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:37 crc kubenswrapper[4828]: I1210 18:56:37.967459 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:37Z","lastTransitionTime":"2025-12-10T18:56:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.070059 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.070127 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.070145 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.070168 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.070184 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.173259 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.173327 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.173353 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.173382 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.173398 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.276681 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.276760 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.276777 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.276844 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.276864 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.379674 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.379732 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.379750 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.379777 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.379795 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.482628 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.482712 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.482736 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.482768 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.482790 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.586364 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.586453 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.586472 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.586503 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.586523 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.690203 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.690288 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.690308 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.690338 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.690362 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.787908 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.787969 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.787927 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.787917 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:38 crc kubenswrapper[4828]: E1210 18:56:38.788218 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:38 crc kubenswrapper[4828]: E1210 18:56:38.788305 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:38 crc kubenswrapper[4828]: E1210 18:56:38.788458 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:38 crc kubenswrapper[4828]: E1210 18:56:38.788600 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.789926 4828 scope.go:117] "RemoveContainer" containerID="fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.795038 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.795076 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.795118 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.795184 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.795211 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.897680 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.898209 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.898224 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.898258 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:38 crc kubenswrapper[4828]: I1210 18:56:38.898268 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:38Z","lastTransitionTime":"2025-12-10T18:56:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.000552 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.000599 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.000611 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.000630 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.000643 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.103684 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.103716 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.103725 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.103741 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.103754 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.206464 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.206532 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.206548 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.206568 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.206582 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.308779 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.308872 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.308890 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.308914 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.308931 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.412470 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.412512 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.412524 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.412540 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.412551 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.516171 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.516280 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.516299 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.516324 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.516343 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.619392 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.619476 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.619499 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.619529 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.619554 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.722118 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.722154 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.722163 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.722179 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.722191 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.811678 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.825306 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.825363 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.825378 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.825404 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.825421 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.829079 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.843299 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.874706 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe415208b7c4a043d3a9466d81931a16b2d600a3b5605eb89c7d2c2ffb1b71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:25Z\\\",\\\"message\\\":\\\"2025-12-10T18:55:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c10c5c27-b43d-4d94-9259-eaa0eddec29a\\\\n2025-12-10T18:55:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c10c5c27-b43d-4d94-9259-eaa0eddec29a to /host/opt/cni/bin/\\\\n2025-12-10T18:55:40Z [verbose] multus-daemon started\\\\n2025-12-10T18:55:40Z [verbose] Readiness Indicator file check\\\\n2025-12-10T18:56:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.888595 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.899962 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f3052a777d2db499f91f40a7517c0f2cd48191aaf6ebc8b17f1c114c2623fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41e994b69b7552394d02f9225cf931743737e1bfcd0448aff5f860ab27afa00b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.917275 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.927371 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.927399 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.927407 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.927421 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.927430 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:39Z","lastTransitionTime":"2025-12-10T18:56:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.932900 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.950079 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.971783 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:06Z\\\",\\\"message\\\":\\\"mplate:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.153\\\\\\\", Port:5443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1210 18:56:06.639052 6459 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z]\\\\nI1210 18:56:06.638925 6459 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-scheduler-operator for network=default : 3.861754ms\\\\nI1210 18:56:06.639267 6459 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.983900 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-74lmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"472ff1d2-0936-4ee0-b738-dc26e2c03559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-74lmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:39 crc kubenswrapper[4828]: I1210 18:56:39.997570 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1c81c13-3c65-4b70-b50a-177b5d07c0a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3f41b839ee9d8b2b5083bd8b5ba463e892011fcfb9052449f9198c6bba5d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://215812f5a8013b31678b85a9fa5232c2da907e04f587be31f9508ea39182976b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://215812f5a8013b31678b85a9fa5232c2da907e04f587be31f9508ea39182976b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:39Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.009370 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.020074 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.029967 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.030011 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.030023 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.030040 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.030051 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.030715 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60cfa1f-f033-4ea6-8785-b00708e0af0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1dcfb31658e3faf516dbebd1e00d1f8dc701e5c6b73831a52648effb712db12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66fc64d5883f4b746a0636317bd472362758bdbfa1c5ab6a0a327692256e064a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b044808efab4ceca84bf0256a92f6559ea57b15a0d2a6c83756f489accbefdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.044253 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.055647 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.066724 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.132357 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.132396 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.132406 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.132426 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.132437 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.219507 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovnkube-controller/2.log" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.222569 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerStarted","Data":"7e0ef04a88803858aabdbc8c1b702a19cc044e5df4e33e4d1a719f4d506e7c82"} Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.223099 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.234695 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.234769 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.234790 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.234859 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.234880 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.242215 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60cfa1f-f033-4ea6-8785-b00708e0af0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1dcfb31658e3faf516dbebd1e00d1f8dc701e5c6b73831a52648effb712db12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66fc64d5883f4b746a0636317bd472362758bdbfa1c5ab6a0a327692256e064a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b044808efab4ceca84bf0256a92f6559ea57b15a0d2a6c83756f489accbefdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.255851 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.269737 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.286417 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.314891 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.331972 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.337839 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.337888 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.337899 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.337917 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.337927 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.343941 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.356967 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe415208b7c4a043d3a9466d81931a16b2d600a3b5605eb89c7d2c2ffb1b71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:25Z\\\",\\\"message\\\":\\\"2025-12-10T18:55:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c10c5c27-b43d-4d94-9259-eaa0eddec29a\\\\n2025-12-10T18:55:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c10c5c27-b43d-4d94-9259-eaa0eddec29a to /host/opt/cni/bin/\\\\n2025-12-10T18:55:40Z [verbose] multus-daemon started\\\\n2025-12-10T18:55:40Z [verbose] Readiness Indicator file check\\\\n2025-12-10T18:56:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.365418 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.376745 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f3052a777d2db499f91f40a7517c0f2cd48191aaf6ebc8b17f1c114c2623fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41e994b69b7552394d02f9225cf931743737e1bfcd0448aff5f860ab27afa00b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.389562 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.403652 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.418554 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.439228 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0ef04a88803858aabdbc8c1b702a19cc044e5df4e33e4d1a719f4d506e7c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:06Z\\\",\\\"message\\\":\\\"mplate:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.153\\\\\\\", Port:5443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1210 18:56:06.639052 6459 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z]\\\\nI1210 18:56:06.638925 6459 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-scheduler-operator for network=default : 3.861754ms\\\\nI1210 18:56:06.639267 6459 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.440124 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.440157 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.440167 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.440181 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.440193 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.450912 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-74lmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"472ff1d2-0936-4ee0-b738-dc26e2c03559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-74lmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.461762 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1c81c13-3c65-4b70-b50a-177b5d07c0a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3f41b839ee9d8b2b5083bd8b5ba463e892011fcfb9052449f9198c6bba5d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://215812f5a8013b31678b85a9fa5232c2da907e04f587be31f9508ea39182976b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://215812f5a8013b31678b85a9fa5232c2da907e04f587be31f9508ea39182976b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.474746 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.485288 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:40Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.542219 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.542270 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.542286 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.542306 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.542320 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.645022 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.645091 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.645112 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.645139 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.645157 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.747109 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.747163 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.747178 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.747197 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.747213 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.788780 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.788818 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.788883 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.788925 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:40 crc kubenswrapper[4828]: E1210 18:56:40.789110 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:40 crc kubenswrapper[4828]: E1210 18:56:40.789166 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:40 crc kubenswrapper[4828]: E1210 18:56:40.789239 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:40 crc kubenswrapper[4828]: E1210 18:56:40.789344 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.849856 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.849911 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.849929 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.849953 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.849973 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.952565 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.952622 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.952630 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.952645 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:40 crc kubenswrapper[4828]: I1210 18:56:40.952654 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:40Z","lastTransitionTime":"2025-12-10T18:56:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.055385 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.055423 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.055436 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.055459 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.055671 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.158319 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.158408 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.158446 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.158466 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.158478 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.241488 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovnkube-controller/3.log" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.243105 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovnkube-controller/2.log" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.247859 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerID="7e0ef04a88803858aabdbc8c1b702a19cc044e5df4e33e4d1a719f4d506e7c82" exitCode=1 Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.247922 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerDied","Data":"7e0ef04a88803858aabdbc8c1b702a19cc044e5df4e33e4d1a719f4d506e7c82"} Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.247988 4828 scope.go:117] "RemoveContainer" containerID="fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.248944 4828 scope.go:117] "RemoveContainer" containerID="7e0ef04a88803858aabdbc8c1b702a19cc044e5df4e33e4d1a719f4d506e7c82" Dec 10 18:56:41 crc kubenswrapper[4828]: E1210 18:56:41.249208 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.262168 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.262224 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.262241 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.262264 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.262283 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.270191 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60cfa1f-f033-4ea6-8785-b00708e0af0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1dcfb31658e3faf516dbebd1e00d1f8dc701e5c6b73831a52648effb712db12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66fc64d5883f4b746a0636317bd472362758bdbfa1c5ab6a0a327692256e064a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b044808efab4ceca84bf0256a92f6559ea57b15a0d2a6c83756f489accbefdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.289469 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.309547 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.323882 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.342137 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.359785 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.365150 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.365212 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.365231 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.365259 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.365278 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.381016 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe415208b7c4a043d3a9466d81931a16b2d600a3b5605eb89c7d2c2ffb1b71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:25Z\\\",\\\"message\\\":\\\"2025-12-10T18:55:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c10c5c27-b43d-4d94-9259-eaa0eddec29a\\\\n2025-12-10T18:55:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c10c5c27-b43d-4d94-9259-eaa0eddec29a to /host/opt/cni/bin/\\\\n2025-12-10T18:55:40Z [verbose] multus-daemon started\\\\n2025-12-10T18:55:40Z [verbose] Readiness Indicator file check\\\\n2025-12-10T18:56:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.394280 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.409881 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f3052a777d2db499f91f40a7517c0f2cd48191aaf6ebc8b17f1c114c2623fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41e994b69b7552394d02f9225cf931743737e1bfcd0448aff5f860ab27afa00b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.424277 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.439430 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.455592 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.468408 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.468464 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.468479 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.468504 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.468521 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.479356 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0ef04a88803858aabdbc8c1b702a19cc044e5df4e33e4d1a719f4d506e7c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa72c7793b1effe8f1f4ec71a24878d561e3b83253696090ca5bc1a0e26fbc00\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:06Z\\\",\\\"message\\\":\\\"mplate:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.153\\\\\\\", Port:5443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1210 18:56:06.639052 6459 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:06Z is after 2025-08-24T17:21:41Z]\\\\nI1210 18:56:06.638925 6459 services_controller.go:360] Finished syncing service metrics on namespace openshift-kube-scheduler-operator for network=default : 3.861754ms\\\\nI1210 18:56:06.639267 6459 model_clien\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e0ef04a88803858aabdbc8c1b702a19cc044e5df4e33e4d1a719f4d506e7c82\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"ap[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-authentication-operator/metrics\\\\\\\"}\\\\nI1210 18:56:40.691190 6863 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1210 18:56:40.691201 6863 services_controller.go:360] Finished syncing service metrics on namespace openshift-authentication-operator for network=default : 2.566292ms\\\\nI1210 18:56:40.691247 6863 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}\\\\nI1210 18:56:40.691278 6863 services_controller.go:360] Finished syncing service metrics on namespace openshift-config-operator for network=default : 2.440398ms\\\\nI1210 18:56:40.691255 6863 factory.go:1336] Added *v1.Node event handler 7\\\\nI1210 18:56:40.691344 6863 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1210 18:56:40.691736 6863 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1210 18:56:40.691878 6863 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1210 18:56:40.691920 6863 ovnkube.go:599] Stopped ovnkube\\\\nI1210 18:56:40.691955 6863 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 18:56:40.692078 6863 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.492597 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-74lmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"472ff1d2-0936-4ee0-b738-dc26e2c03559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-74lmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.510843 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.525092 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.538708 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.553174 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1c81c13-3c65-4b70-b50a-177b5d07c0a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3f41b839ee9d8b2b5083bd8b5ba463e892011fcfb9052449f9198c6bba5d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://215812f5a8013b31678b85a9fa5232c2da907e04f587be31f9508ea39182976b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://215812f5a8013b31678b85a9fa5232c2da907e04f587be31f9508ea39182976b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:41Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.571038 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.571072 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.571082 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.571095 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.571104 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.673744 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.673853 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.673883 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.673917 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.673941 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.777077 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.777127 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.777144 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.777168 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.777187 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.879851 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.879891 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.879902 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.879921 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.879933 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.983485 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.983549 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.983566 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.983591 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:41 crc kubenswrapper[4828]: I1210 18:56:41.983611 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:41Z","lastTransitionTime":"2025-12-10T18:56:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.086884 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.086939 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.086953 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.086976 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.086988 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.190693 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.190758 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.190773 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.190789 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.190814 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.253521 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovnkube-controller/3.log" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.256786 4828 scope.go:117] "RemoveContainer" containerID="7e0ef04a88803858aabdbc8c1b702a19cc044e5df4e33e4d1a719f4d506e7c82" Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.257055 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.271287 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fb4e3e43-3654-4f02-9ef7-d9ab9e666da2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e3f3052a777d2db499f91f40a7517c0f2cd48191aaf6ebc8b17f1c114c2623fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41e994b69b7552394d02f9225cf931743737e1bfcd0448aff5f860ab27afa00b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kw5sr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-pwjml\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.285865 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.294069 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.294139 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.294168 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.294205 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.294232 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.303423 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49d1c3dd40172f8d6c440ff839e0872f6f69f46e8b7e28c54185dc88c60267d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.315844 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://15b239965f1c5ed7421d22ea2349d41812105d792c5757c198e3474f6cde7f60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qp2lf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-677ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.329606 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xdxgd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c6e55203-65ae-4767-9a96-f309bc41ddf3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe415208b7c4a043d3a9466d81931a16b2d600a3b5605eb89c7d2c2ffb1b71b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:25Z\\\",\\\"message\\\":\\\"2025-12-10T18:55:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c10c5c27-b43d-4d94-9259-eaa0eddec29a\\\\n2025-12-10T18:55:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c10c5c27-b43d-4d94-9259-eaa0eddec29a to /host/opt/cni/bin/\\\\n2025-12-10T18:55:40Z [verbose] multus-daemon started\\\\n2025-12-10T18:55:40Z [verbose] Readiness Indicator file check\\\\n2025-12-10T18:56:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:56:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-srfr2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xdxgd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.343201 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5px8v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ff6072ff-b4d5-48a1-9ee5-03458bcbd758\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95fcf697397993318d2d87fe3cf330e941a86510d38ddd08a1d492e3ae054073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-84wpn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:40Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5px8v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.356321 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fe272100-65b7-4be3-9ab4-51e4a6f475de\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1210 18:55:32.087263 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1210 18:55:32.089312 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3026580255/tls.crt::/tmp/serving-cert-3026580255/tls.key\\\\\\\"\\\\nI1210 18:55:37.900822 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1210 18:55:37.906410 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1210 18:55:37.906547 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1210 18:55:37.906638 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1210 18:55:37.906685 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1210 18:55:37.917870 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1210 18:55:37.917918 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1210 18:55:37.917929 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918043 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1210 18:55:37.918055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1210 18:55:37.918064 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1210 18:55:37.918073 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1210 18:55:37.918083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1210 18:55:37.922554 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.373231 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5aad9ab1-5050-4e35-bae1-6cef7b26da53\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a409ebecbaef05ccbe12d2e687d65ea798923770a37bdfa796ba85d37a305a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a36e93429e06f84ec31e78eaf962b0ca58f013f5589a5e4761621f2302a719e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://345c3ffc2cc30f913617165f26cb78b851b6bb3c01ed19c2b0c6f342ecbb6a3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.397761 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rzndz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40784d67-bada-4229-8764-efba7c96edb5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cd649337ada44a83922268371d2609a5e1fd145f514536202c3791b180a22587\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://039b42ae4f817a7b8b2f7e601a2060f42028324fdc58a6a0d5ce0fd8b5645b06\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://55170086e72db3c1e0f779c9f913a81a83d9314e9f7a4273a51b6a3d236d6d0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://acfb1737eb52f169d877fa4bc132fd6783f051fe6b3fb9e37fd2d14e16e561cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d16cb634971c7c90bc0af4f8e664ddde9d12ca0af32eff6964bd91508722ff2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988824c3b703c6894843ee6ffc3ca52fd2f1f08dc0bda6a1eccc2e0af6e216f9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7ad6fff6644117da7d87b5c24a65eb90d53682102a28e502b8cecd8f9636b58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wsnvv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rzndz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.398840 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.398922 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.398943 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.398973 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.398999 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.433348 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f37f60f-fe12-4530-8ff1-ea3794c65597\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e0ef04a88803858aabdbc8c1b702a19cc044e5df4e33e4d1a719f4d506e7c82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e0ef04a88803858aabdbc8c1b702a19cc044e5df4e33e4d1a719f4d506e7c82\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-10T18:56:40Z\\\",\\\"message\\\":\\\"ap[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-authentication-operator/metrics\\\\\\\"}\\\\nI1210 18:56:40.691190 6863 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1210 18:56:40.691201 6863 services_controller.go:360] Finished syncing service metrics on namespace openshift-authentication-operator for network=default : 2.566292ms\\\\nI1210 18:56:40.691247 6863 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-config-operator/metrics\\\\\\\"}\\\\nI1210 18:56:40.691278 6863 services_controller.go:360] Finished syncing service metrics on namespace openshift-config-operator for network=default : 2.440398ms\\\\nI1210 18:56:40.691255 6863 factory.go:1336] Added *v1.Node event handler 7\\\\nI1210 18:56:40.691344 6863 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1210 18:56:40.691736 6863 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1210 18:56:40.691878 6863 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1210 18:56:40.691920 6863 ovnkube.go:599] Stopped ovnkube\\\\nI1210 18:56:40.691955 6863 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1210 18:56:40.692078 6863 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-10T18:56:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4hdv4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8f42x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.456647 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-74lmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"472ff1d2-0936-4ee0-b738-dc26e2c03559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:53Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6zq55\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:53Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-74lmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.474233 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1c81c13-3c65-4b70-b50a-177b5d07c0a7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f3f41b839ee9d8b2b5083bd8b5ba463e892011fcfb9052449f9198c6bba5d64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://215812f5a8013b31678b85a9fa5232c2da907e04f587be31f9508ea39182976b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://215812f5a8013b31678b85a9fa5232c2da907e04f587be31f9508ea39182976b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.494526 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b4813a7b06fa02e3693ec8dcab7a184d7ca8ac56034a01d8737948072f79fc11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.502183 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.502288 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.502351 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.502424 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.502492 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.517672 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.538251 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d60cfa1f-f033-4ea6-8785-b00708e0af0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1dcfb31658e3faf516dbebd1e00d1f8dc701e5c6b73831a52648effb712db12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66fc64d5883f4b746a0636317bd472362758bdbfa1c5ab6a0a327692256e064a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b044808efab4ceca84bf0256a92f6559ea57b15a0d2a6c83756f489accbefdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61b74b07762aac4e56bb0adba60b09e14ea773640f5cfcb454486760a682b8f6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-10T18:55:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-10T18:55:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.558621 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.577094 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2c2b7846a265988b79fee83f1568b421e34b36c6b0688c2d0f30bf50f041551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13dfe63e38e253d0ab58a8f37a4588a5218759a14f256dea1b8a5adbc2ef0d89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.596546 4828 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-hnf4q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"74b63880-331e-46a7-895b-434b52ff35b2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-10T18:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5de769aa1f68c9b67fe2a1c404b2949916a41732b0421ebf089c1819c238593f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-10T18:55:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mtbf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-10T18:55:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-hnf4q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:42Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.605586 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.605650 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.605668 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.605694 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.605716 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.709839 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.709912 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.709931 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.709956 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.709973 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.720279 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.720444 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:46.720408174 +0000 UTC m=+147.231019229 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.720511 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.720678 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.720727 4828 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.720865 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:46.720839597 +0000 UTC m=+147.231450652 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.720936 4828 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.721084 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:46.721038382 +0000 UTC m=+147.231649427 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.788372 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.788423 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.788458 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.788591 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.788391 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.788901 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.789034 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.789228 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.815673 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.815727 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.815745 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.815773 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.815795 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.822510 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.822570 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.822760 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.822782 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.822880 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.822901 4828 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.822833 4828 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.822985 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:46.82295986 +0000 UTC m=+147.333570895 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.823003 4828 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:56:42 crc kubenswrapper[4828]: E1210 18:56:42.823076 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-10 18:57:46.823053873 +0000 UTC m=+147.333664908 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.919577 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.919649 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.919668 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.919700 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:42 crc kubenswrapper[4828]: I1210 18:56:42.919717 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:42Z","lastTransitionTime":"2025-12-10T18:56:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.023103 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.023171 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.023188 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.023215 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.023234 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.126705 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.126762 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.126778 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.126830 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.126849 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.230057 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.230117 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.230134 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.230159 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.230176 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.333329 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.333389 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.333406 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.333433 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.333515 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.436992 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.437087 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.437136 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.437179 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.437196 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.454996 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.455067 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.455096 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.455131 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.455156 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4828]: E1210 18:56:43.476407 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.481894 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.481977 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.482006 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.482038 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.482060 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4828]: E1210 18:56:43.508847 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.515317 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.515375 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.515393 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.515420 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.515437 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4828]: E1210 18:56:43.539385 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.545962 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.546033 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.546057 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.546592 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.546656 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4828]: E1210 18:56:43.570360 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.575125 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.575192 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.575222 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.575288 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.575321 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4828]: E1210 18:56:43.595861 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-10T18:56:43Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"df53ff5c-912c-4420-b6db-884332bb2d58\\\",\\\"systemUUID\\\":\\\"2c4a1f25-f3cd-4108-b88c-83e56814cf09\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-10T18:56:43Z is after 2025-08-24T17:21:41Z" Dec 10 18:56:43 crc kubenswrapper[4828]: E1210 18:56:43.596448 4828 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.599092 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.599378 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.599595 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.599772 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.599959 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.703553 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.703973 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.704127 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.704342 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.704488 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.807271 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.807350 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.807378 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.807410 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.807433 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.912082 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.912140 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.912158 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.912301 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:43 crc kubenswrapper[4828]: I1210 18:56:43.912554 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:43Z","lastTransitionTime":"2025-12-10T18:56:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.015304 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.015371 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.015387 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.015411 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.015431 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.119006 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.119087 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.119113 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.119137 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.119156 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.226683 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.226770 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.226835 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.226876 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.226907 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.329545 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.329916 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.330073 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.330210 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.330346 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.434130 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.434190 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.434208 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.434237 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.434255 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.537477 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.537566 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.537596 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.537630 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.537654 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.640106 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.640144 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.640156 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.640173 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.640184 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.742938 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.742994 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.743006 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.743023 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.743036 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.787958 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.788073 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.787971 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:44 crc kubenswrapper[4828]: E1210 18:56:44.788153 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.787995 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:44 crc kubenswrapper[4828]: E1210 18:56:44.788263 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:44 crc kubenswrapper[4828]: E1210 18:56:44.788443 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:44 crc kubenswrapper[4828]: E1210 18:56:44.788488 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.846632 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.846704 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.846727 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.846756 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.846777 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.949988 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.950058 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.950075 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.950100 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:44 crc kubenswrapper[4828]: I1210 18:56:44.950121 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:44Z","lastTransitionTime":"2025-12-10T18:56:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.053278 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.053376 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.053396 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.053425 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.053442 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.156295 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.156376 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.156395 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.156422 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.156443 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.259419 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.259761 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.260060 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.260216 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.260365 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.364075 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.364133 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.364153 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.364179 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.364197 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.467123 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.467502 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.467666 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.467892 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.468100 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.571875 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.572442 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.572689 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.572953 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.573111 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.676643 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.677311 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.677433 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.677521 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.677600 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.780579 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.780879 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.781007 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.781172 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.781281 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.883596 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.883945 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.884107 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.884243 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.884377 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.987602 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.987899 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.987971 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.988046 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:45 crc kubenswrapper[4828]: I1210 18:56:45.988115 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:45Z","lastTransitionTime":"2025-12-10T18:56:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.091369 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.091428 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.091446 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.091470 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.091490 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.194532 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.194584 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.194598 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.194618 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.194632 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.297959 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.298017 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.298033 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.298060 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.298077 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.401153 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.401240 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.401262 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.401294 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.401317 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.505152 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.505208 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.505221 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.505241 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.505261 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.608871 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.608972 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.609006 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.609055 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.609078 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.711939 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.711984 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.711998 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.712016 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.712028 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.788036 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.788103 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.788144 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.788176 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:46 crc kubenswrapper[4828]: E1210 18:56:46.788168 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:46 crc kubenswrapper[4828]: E1210 18:56:46.788282 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:46 crc kubenswrapper[4828]: E1210 18:56:46.788396 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:46 crc kubenswrapper[4828]: E1210 18:56:46.788446 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.814202 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.814225 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.814233 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.814243 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.814251 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.917310 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.917386 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.917410 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.917458 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:46 crc kubenswrapper[4828]: I1210 18:56:46.917483 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:46Z","lastTransitionTime":"2025-12-10T18:56:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.020283 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.020333 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.020347 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.020366 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.020382 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.122531 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.122575 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.122588 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.122608 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.122623 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.225840 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.225891 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.225903 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.225923 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.225940 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.328315 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.328369 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.328382 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.328401 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.328414 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.430919 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.430966 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.430978 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.430996 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.431007 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.534136 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.534196 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.534211 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.534233 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.534247 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.637501 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.637538 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.637549 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.637565 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.637577 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.740577 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.740626 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.740639 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.740657 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.740670 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.843272 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.843332 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.843343 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.843359 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.843369 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.945693 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.945751 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.945759 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.945775 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:47 crc kubenswrapper[4828]: I1210 18:56:47.945785 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:47Z","lastTransitionTime":"2025-12-10T18:56:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.048839 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.048903 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.048920 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.048946 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.048963 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.151299 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.151369 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.151387 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.151413 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.151429 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.253391 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.253453 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.253463 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.253479 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.253488 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.355853 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.355972 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.355987 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.356008 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.356019 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.458741 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.458823 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.458841 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.458860 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.458874 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.561654 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.561689 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.561697 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.561714 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.561724 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.664581 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.664649 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.664672 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.664704 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.664727 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.766856 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.766907 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.766923 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.766945 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.766964 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.788568 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.788618 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.788620 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.788675 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:48 crc kubenswrapper[4828]: E1210 18:56:48.788836 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:48 crc kubenswrapper[4828]: E1210 18:56:48.788925 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:48 crc kubenswrapper[4828]: E1210 18:56:48.789146 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:48 crc kubenswrapper[4828]: E1210 18:56:48.789189 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.869408 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.869449 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.869459 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.869476 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.869486 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.972194 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.972234 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.972246 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.972260 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:48 crc kubenswrapper[4828]: I1210 18:56:48.972269 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:48Z","lastTransitionTime":"2025-12-10T18:56:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.074990 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.075042 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.075054 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.075072 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.075086 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.178480 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.178568 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.178586 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.178612 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.178630 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.280584 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.280650 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.280659 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.280674 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.280684 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.382770 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.382826 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.382836 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.382851 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.382863 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.484682 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.484727 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.484741 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.484788 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.484815 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.587496 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.587549 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.587566 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.587590 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.587607 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.690626 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.690688 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.690705 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.690732 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.690753 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.799345 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.799424 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.799444 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.799470 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.799497 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.820626 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-rzndz" podStartSLOduration=71.820602515 podStartE2EDuration="1m11.820602515s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:49.820251245 +0000 UTC m=+90.330862330" watchObservedRunningTime="2025-12-10 18:56:49.820602515 +0000 UTC m=+90.331213550" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.891348 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=71.891323925 podStartE2EDuration="1m11.891323925s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:49.891002926 +0000 UTC m=+90.401613941" watchObservedRunningTime="2025-12-10 18:56:49.891323925 +0000 UTC m=+90.401934950" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.901831 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.902137 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.902266 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.902425 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.902552 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:49Z","lastTransitionTime":"2025-12-10T18:56:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.905152 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=72.905131584 podStartE2EDuration="1m12.905131584s" podCreationTimestamp="2025-12-10 18:55:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:49.904468785 +0000 UTC m=+90.415079780" watchObservedRunningTime="2025-12-10 18:56:49.905131584 +0000 UTC m=+90.415742609" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.941733 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=27.941708033 podStartE2EDuration="27.941708033s" podCreationTimestamp="2025-12-10 18:56:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:49.928624285 +0000 UTC m=+90.439235290" watchObservedRunningTime="2025-12-10 18:56:49.941708033 +0000 UTC m=+90.452319048" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.975379 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-hnf4q" podStartSLOduration=71.97535529 podStartE2EDuration="1m11.97535529s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:49.974115515 +0000 UTC m=+90.484726520" watchObservedRunningTime="2025-12-10 18:56:49.97535529 +0000 UTC m=+90.485966295" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.997442 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podStartSLOduration=71.997421691 podStartE2EDuration="1m11.997421691s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:49.997342589 +0000 UTC m=+90.507953594" watchObservedRunningTime="2025-12-10 18:56:49.997421691 +0000 UTC m=+90.508032696" Dec 10 18:56:49 crc kubenswrapper[4828]: I1210 18:56:49.997539 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=38.997536064 podStartE2EDuration="38.997536064s" podCreationTimestamp="2025-12-10 18:56:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:49.985857996 +0000 UTC m=+90.496469011" watchObservedRunningTime="2025-12-10 18:56:49.997536064 +0000 UTC m=+90.508147069" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.004639 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.004841 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.004948 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.005054 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.005180 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.020114 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-xdxgd" podStartSLOduration=72.020098919 podStartE2EDuration="1m12.020098919s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:50.012196317 +0000 UTC m=+90.522807312" watchObservedRunningTime="2025-12-10 18:56:50.020098919 +0000 UTC m=+90.530709924" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.032169 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-5px8v" podStartSLOduration=72.032149038 podStartE2EDuration="1m12.032149038s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:50.020653185 +0000 UTC m=+90.531264180" watchObservedRunningTime="2025-12-10 18:56:50.032149038 +0000 UTC m=+90.542760043" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.046036 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-pwjml" podStartSLOduration=71.046011849 podStartE2EDuration="1m11.046011849s" podCreationTimestamp="2025-12-10 18:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:50.033648451 +0000 UTC m=+90.544259456" watchObservedRunningTime="2025-12-10 18:56:50.046011849 +0000 UTC m=+90.556622854" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.107389 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.107431 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.107443 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.107464 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.107476 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.209857 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.210215 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.210391 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.210565 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.210767 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.314023 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.314079 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.314089 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.314107 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.314118 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.416314 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.416350 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.416359 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.416373 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.416383 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.519234 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.519305 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.519322 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.519347 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.519365 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.621836 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.621872 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.621881 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.621900 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.621915 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.724817 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.724854 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.724862 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.724876 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.724885 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.788392 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.788439 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.788481 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:50 crc kubenswrapper[4828]: E1210 18:56:50.788579 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.788646 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:50 crc kubenswrapper[4828]: E1210 18:56:50.788757 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:50 crc kubenswrapper[4828]: E1210 18:56:50.788893 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:50 crc kubenswrapper[4828]: E1210 18:56:50.788994 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.827118 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.827154 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.827167 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.827184 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.827194 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.929491 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.929785 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.929985 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.930140 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:50 crc kubenswrapper[4828]: I1210 18:56:50.930231 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:50Z","lastTransitionTime":"2025-12-10T18:56:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.032607 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.032664 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.032677 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.032697 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.032710 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:51Z","lastTransitionTime":"2025-12-10T18:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.135321 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.135364 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.135374 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.135400 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.135413 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:51Z","lastTransitionTime":"2025-12-10T18:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.237645 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.237723 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.237751 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.237790 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.237863 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:51Z","lastTransitionTime":"2025-12-10T18:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.340411 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.340467 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.340483 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.340507 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.340526 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:51Z","lastTransitionTime":"2025-12-10T18:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.443171 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.443225 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.443239 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.443257 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.443269 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:51Z","lastTransitionTime":"2025-12-10T18:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.545444 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.545487 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.545499 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.545516 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.545529 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:51Z","lastTransitionTime":"2025-12-10T18:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.647928 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.648009 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.648025 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.648048 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.648064 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:51Z","lastTransitionTime":"2025-12-10T18:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.750892 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.750971 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.750990 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.751017 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.751033 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:51Z","lastTransitionTime":"2025-12-10T18:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.853703 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.853782 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.853813 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.853837 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.853862 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:51Z","lastTransitionTime":"2025-12-10T18:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.956285 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.956322 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.956335 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.956352 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:51 crc kubenswrapper[4828]: I1210 18:56:51.956364 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:51Z","lastTransitionTime":"2025-12-10T18:56:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.060507 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.060575 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.060594 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.060620 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.060637 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:52Z","lastTransitionTime":"2025-12-10T18:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.163489 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.163561 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.163572 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.163591 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.163602 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:52Z","lastTransitionTime":"2025-12-10T18:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.265699 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.265742 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.265750 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.265766 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.265774 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:52Z","lastTransitionTime":"2025-12-10T18:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.368465 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.368518 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.368531 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.368551 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.368563 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:52Z","lastTransitionTime":"2025-12-10T18:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.471356 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.471415 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.471424 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.471440 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.471449 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:52Z","lastTransitionTime":"2025-12-10T18:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.574443 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.574481 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.574489 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.574504 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.574513 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:52Z","lastTransitionTime":"2025-12-10T18:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.677561 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.677609 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.677622 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.677640 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.677653 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:52Z","lastTransitionTime":"2025-12-10T18:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.781064 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.781133 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.781152 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.781178 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.781196 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:52Z","lastTransitionTime":"2025-12-10T18:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.788367 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.788424 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.788456 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.788486 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:52 crc kubenswrapper[4828]: E1210 18:56:52.788548 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:52 crc kubenswrapper[4828]: E1210 18:56:52.788664 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:52 crc kubenswrapper[4828]: E1210 18:56:52.789011 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:52 crc kubenswrapper[4828]: E1210 18:56:52.789067 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.883865 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.883911 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.883922 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.883938 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.883949 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:52Z","lastTransitionTime":"2025-12-10T18:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.986411 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.986465 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.986478 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.986497 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:52 crc kubenswrapper[4828]: I1210 18:56:52.986509 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:52Z","lastTransitionTime":"2025-12-10T18:56:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.090008 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.090084 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.090102 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.090129 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.090148 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:53Z","lastTransitionTime":"2025-12-10T18:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.192534 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.192625 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.192653 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.192691 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.192715 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:53Z","lastTransitionTime":"2025-12-10T18:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.295927 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.295976 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.295985 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.296004 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.296014 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:53Z","lastTransitionTime":"2025-12-10T18:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.398994 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.399064 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.399085 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.399119 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.399143 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:53Z","lastTransitionTime":"2025-12-10T18:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.502614 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.502669 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.502683 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.502704 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.502718 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:53Z","lastTransitionTime":"2025-12-10T18:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.605377 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.605436 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.605449 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.605467 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.605488 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:53Z","lastTransitionTime":"2025-12-10T18:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.709359 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.709420 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.709432 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.709455 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.709468 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:53Z","lastTransitionTime":"2025-12-10T18:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.784995 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.785086 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.785109 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.785139 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.785162 4828 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-10T18:56:53Z","lastTransitionTime":"2025-12-10T18:56:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.788735 4828 scope.go:117] "RemoveContainer" containerID="7e0ef04a88803858aabdbc8c1b702a19cc044e5df4e33e4d1a719f4d506e7c82" Dec 10 18:56:53 crc kubenswrapper[4828]: E1210 18:56:53.788926 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.840212 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778"] Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.840986 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.843229 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.843378 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.843535 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.843987 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.946420 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/be0fd1b5-2462-4d28-84ab-f6875533beb0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6g778\" (UID: \"be0fd1b5-2462-4d28-84ab-f6875533beb0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.946529 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/be0fd1b5-2462-4d28-84ab-f6875533beb0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6g778\" (UID: \"be0fd1b5-2462-4d28-84ab-f6875533beb0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.946603 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be0fd1b5-2462-4d28-84ab-f6875533beb0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6g778\" (UID: \"be0fd1b5-2462-4d28-84ab-f6875533beb0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.946665 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/be0fd1b5-2462-4d28-84ab-f6875533beb0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6g778\" (UID: \"be0fd1b5-2462-4d28-84ab-f6875533beb0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" Dec 10 18:56:53 crc kubenswrapper[4828]: I1210 18:56:53.946702 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be0fd1b5-2462-4d28-84ab-f6875533beb0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6g778\" (UID: \"be0fd1b5-2462-4d28-84ab-f6875533beb0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" Dec 10 18:56:54 crc kubenswrapper[4828]: I1210 18:56:54.047916 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/be0fd1b5-2462-4d28-84ab-f6875533beb0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6g778\" (UID: \"be0fd1b5-2462-4d28-84ab-f6875533beb0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" Dec 10 18:56:54 crc kubenswrapper[4828]: I1210 18:56:54.047967 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/be0fd1b5-2462-4d28-84ab-f6875533beb0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6g778\" (UID: \"be0fd1b5-2462-4d28-84ab-f6875533beb0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" Dec 10 18:56:54 crc kubenswrapper[4828]: I1210 18:56:54.048101 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/be0fd1b5-2462-4d28-84ab-f6875533beb0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6g778\" (UID: \"be0fd1b5-2462-4d28-84ab-f6875533beb0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" Dec 10 18:56:54 crc kubenswrapper[4828]: I1210 18:56:54.049667 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be0fd1b5-2462-4d28-84ab-f6875533beb0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6g778\" (UID: \"be0fd1b5-2462-4d28-84ab-f6875533beb0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" Dec 10 18:56:54 crc kubenswrapper[4828]: I1210 18:56:54.049708 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/be0fd1b5-2462-4d28-84ab-f6875533beb0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6g778\" (UID: \"be0fd1b5-2462-4d28-84ab-f6875533beb0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" Dec 10 18:56:54 crc kubenswrapper[4828]: I1210 18:56:54.049729 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be0fd1b5-2462-4d28-84ab-f6875533beb0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6g778\" (UID: \"be0fd1b5-2462-4d28-84ab-f6875533beb0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" Dec 10 18:56:54 crc kubenswrapper[4828]: I1210 18:56:54.049837 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/be0fd1b5-2462-4d28-84ab-f6875533beb0-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6g778\" (UID: \"be0fd1b5-2462-4d28-84ab-f6875533beb0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" Dec 10 18:56:54 crc kubenswrapper[4828]: I1210 18:56:54.050840 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/be0fd1b5-2462-4d28-84ab-f6875533beb0-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6g778\" (UID: \"be0fd1b5-2462-4d28-84ab-f6875533beb0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" Dec 10 18:56:54 crc kubenswrapper[4828]: I1210 18:56:54.054873 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be0fd1b5-2462-4d28-84ab-f6875533beb0-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6g778\" (UID: \"be0fd1b5-2462-4d28-84ab-f6875533beb0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" Dec 10 18:56:54 crc kubenswrapper[4828]: I1210 18:56:54.076963 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be0fd1b5-2462-4d28-84ab-f6875533beb0-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6g778\" (UID: \"be0fd1b5-2462-4d28-84ab-f6875533beb0\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" Dec 10 18:56:54 crc kubenswrapper[4828]: I1210 18:56:54.159521 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" Dec 10 18:56:54 crc kubenswrapper[4828]: I1210 18:56:54.300305 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" event={"ID":"be0fd1b5-2462-4d28-84ab-f6875533beb0","Type":"ContainerStarted","Data":"bba3a888ce8c1dbb94092e72daeaa3b9b9ab6b6c0ea5e90889e8246bd79e06cb"} Dec 10 18:56:54 crc kubenswrapper[4828]: I1210 18:56:54.788245 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:54 crc kubenswrapper[4828]: I1210 18:56:54.788356 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:54 crc kubenswrapper[4828]: E1210 18:56:54.788403 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:54 crc kubenswrapper[4828]: I1210 18:56:54.788439 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:54 crc kubenswrapper[4828]: I1210 18:56:54.788453 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:54 crc kubenswrapper[4828]: E1210 18:56:54.788606 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:54 crc kubenswrapper[4828]: E1210 18:56:54.788660 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:54 crc kubenswrapper[4828]: E1210 18:56:54.788717 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:55 crc kubenswrapper[4828]: I1210 18:56:55.304888 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" event={"ID":"be0fd1b5-2462-4d28-84ab-f6875533beb0","Type":"ContainerStarted","Data":"13cdda27aeee59b8d3d180752d1d3d2c804b1750671541ee698694620675967e"} Dec 10 18:56:56 crc kubenswrapper[4828]: I1210 18:56:56.788336 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:56 crc kubenswrapper[4828]: I1210 18:56:56.788384 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:56 crc kubenswrapper[4828]: I1210 18:56:56.788448 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:56 crc kubenswrapper[4828]: E1210 18:56:56.788529 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:56 crc kubenswrapper[4828]: I1210 18:56:56.788552 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:56 crc kubenswrapper[4828]: E1210 18:56:56.788666 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:56:56 crc kubenswrapper[4828]: E1210 18:56:56.788893 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:56 crc kubenswrapper[4828]: E1210 18:56:56.789117 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:56 crc kubenswrapper[4828]: I1210 18:56:56.981673 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs\") pod \"network-metrics-daemon-74lmx\" (UID: \"472ff1d2-0936-4ee0-b738-dc26e2c03559\") " pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:56 crc kubenswrapper[4828]: E1210 18:56:56.981817 4828 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:56 crc kubenswrapper[4828]: E1210 18:56:56.981898 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs podName:472ff1d2-0936-4ee0-b738-dc26e2c03559 nodeName:}" failed. No retries permitted until 2025-12-10 18:58:00.981877674 +0000 UTC m=+161.492488689 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs") pod "network-metrics-daemon-74lmx" (UID: "472ff1d2-0936-4ee0-b738-dc26e2c03559") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 10 18:56:58 crc kubenswrapper[4828]: I1210 18:56:58.788677 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:56:58 crc kubenswrapper[4828]: I1210 18:56:58.788740 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:56:58 crc kubenswrapper[4828]: I1210 18:56:58.788763 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:56:58 crc kubenswrapper[4828]: E1210 18:56:58.788897 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:56:58 crc kubenswrapper[4828]: I1210 18:56:58.788962 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:56:58 crc kubenswrapper[4828]: E1210 18:56:58.789103 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:56:58 crc kubenswrapper[4828]: E1210 18:56:58.789235 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:56:58 crc kubenswrapper[4828]: E1210 18:56:58.789312 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:00 crc kubenswrapper[4828]: I1210 18:57:00.788209 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:00 crc kubenswrapper[4828]: I1210 18:57:00.788286 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:00 crc kubenswrapper[4828]: I1210 18:57:00.788239 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:57:00 crc kubenswrapper[4828]: I1210 18:57:00.788212 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:00 crc kubenswrapper[4828]: E1210 18:57:00.788523 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:00 crc kubenswrapper[4828]: E1210 18:57:00.789027 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:57:00 crc kubenswrapper[4828]: E1210 18:57:00.789100 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:00 crc kubenswrapper[4828]: E1210 18:57:00.789163 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:00 crc kubenswrapper[4828]: I1210 18:57:00.810924 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6g778" podStartSLOduration=82.810902148 podStartE2EDuration="1m22.810902148s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:56:55.325730288 +0000 UTC m=+95.836341293" watchObservedRunningTime="2025-12-10 18:57:00.810902148 +0000 UTC m=+101.321513173" Dec 10 18:57:00 crc kubenswrapper[4828]: I1210 18:57:00.811295 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 10 18:57:02 crc kubenswrapper[4828]: I1210 18:57:02.788588 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:02 crc kubenswrapper[4828]: I1210 18:57:02.788589 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:02 crc kubenswrapper[4828]: I1210 18:57:02.788595 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:57:02 crc kubenswrapper[4828]: I1210 18:57:02.789267 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:02 crc kubenswrapper[4828]: E1210 18:57:02.789425 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:02 crc kubenswrapper[4828]: E1210 18:57:02.789670 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:57:02 crc kubenswrapper[4828]: E1210 18:57:02.789767 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:02 crc kubenswrapper[4828]: E1210 18:57:02.789950 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:04 crc kubenswrapper[4828]: I1210 18:57:04.788280 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:04 crc kubenswrapper[4828]: I1210 18:57:04.788282 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:57:04 crc kubenswrapper[4828]: I1210 18:57:04.788982 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:04 crc kubenswrapper[4828]: E1210 18:57:04.789177 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:04 crc kubenswrapper[4828]: I1210 18:57:04.789339 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:04 crc kubenswrapper[4828]: E1210 18:57:04.790081 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:04 crc kubenswrapper[4828]: E1210 18:57:04.790235 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:57:04 crc kubenswrapper[4828]: E1210 18:57:04.790330 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:06 crc kubenswrapper[4828]: I1210 18:57:06.788758 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:06 crc kubenswrapper[4828]: I1210 18:57:06.788949 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:57:06 crc kubenswrapper[4828]: I1210 18:57:06.789041 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:06 crc kubenswrapper[4828]: I1210 18:57:06.789050 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:06 crc kubenswrapper[4828]: E1210 18:57:06.788965 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:06 crc kubenswrapper[4828]: E1210 18:57:06.789298 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:57:06 crc kubenswrapper[4828]: E1210 18:57:06.790039 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:06 crc kubenswrapper[4828]: E1210 18:57:06.791138 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:08 crc kubenswrapper[4828]: I1210 18:57:08.788520 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:08 crc kubenswrapper[4828]: E1210 18:57:08.788674 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:08 crc kubenswrapper[4828]: I1210 18:57:08.788681 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:08 crc kubenswrapper[4828]: I1210 18:57:08.788768 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:08 crc kubenswrapper[4828]: I1210 18:57:08.788714 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:57:08 crc kubenswrapper[4828]: E1210 18:57:08.788836 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:08 crc kubenswrapper[4828]: E1210 18:57:08.788943 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:08 crc kubenswrapper[4828]: E1210 18:57:08.789123 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:57:08 crc kubenswrapper[4828]: I1210 18:57:08.789911 4828 scope.go:117] "RemoveContainer" containerID="7e0ef04a88803858aabdbc8c1b702a19cc044e5df4e33e4d1a719f4d506e7c82" Dec 10 18:57:08 crc kubenswrapper[4828]: E1210 18:57:08.790038 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8f42x_openshift-ovn-kubernetes(1f37f60f-fe12-4530-8ff1-ea3794c65597)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" Dec 10 18:57:09 crc kubenswrapper[4828]: I1210 18:57:09.824673 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=9.824639297000001 podStartE2EDuration="9.824639297s" podCreationTimestamp="2025-12-10 18:57:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:09.822434435 +0000 UTC m=+110.333045440" watchObservedRunningTime="2025-12-10 18:57:09.824639297 +0000 UTC m=+110.335250352" Dec 10 18:57:10 crc kubenswrapper[4828]: I1210 18:57:10.788323 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:10 crc kubenswrapper[4828]: I1210 18:57:10.788442 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:10 crc kubenswrapper[4828]: E1210 18:57:10.788486 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:10 crc kubenswrapper[4828]: I1210 18:57:10.788329 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:57:10 crc kubenswrapper[4828]: E1210 18:57:10.788680 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:10 crc kubenswrapper[4828]: E1210 18:57:10.788761 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:57:10 crc kubenswrapper[4828]: I1210 18:57:10.788350 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:10 crc kubenswrapper[4828]: E1210 18:57:10.788861 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:12 crc kubenswrapper[4828]: I1210 18:57:12.366376 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xdxgd_c6e55203-65ae-4767-9a96-f309bc41ddf3/kube-multus/1.log" Dec 10 18:57:12 crc kubenswrapper[4828]: I1210 18:57:12.367300 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xdxgd_c6e55203-65ae-4767-9a96-f309bc41ddf3/kube-multus/0.log" Dec 10 18:57:12 crc kubenswrapper[4828]: I1210 18:57:12.367351 4828 generic.go:334] "Generic (PLEG): container finished" podID="c6e55203-65ae-4767-9a96-f309bc41ddf3" containerID="bbe415208b7c4a043d3a9466d81931a16b2d600a3b5605eb89c7d2c2ffb1b71b" exitCode=1 Dec 10 18:57:12 crc kubenswrapper[4828]: I1210 18:57:12.367388 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xdxgd" event={"ID":"c6e55203-65ae-4767-9a96-f309bc41ddf3","Type":"ContainerDied","Data":"bbe415208b7c4a043d3a9466d81931a16b2d600a3b5605eb89c7d2c2ffb1b71b"} Dec 10 18:57:12 crc kubenswrapper[4828]: I1210 18:57:12.367434 4828 scope.go:117] "RemoveContainer" containerID="6e927b0d6c7b17961791d943c4115975c099adc35586cba49f179085fc0ea085" Dec 10 18:57:12 crc kubenswrapper[4828]: I1210 18:57:12.369099 4828 scope.go:117] "RemoveContainer" containerID="bbe415208b7c4a043d3a9466d81931a16b2d600a3b5605eb89c7d2c2ffb1b71b" Dec 10 18:57:12 crc kubenswrapper[4828]: E1210 18:57:12.369434 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-xdxgd_openshift-multus(c6e55203-65ae-4767-9a96-f309bc41ddf3)\"" pod="openshift-multus/multus-xdxgd" podUID="c6e55203-65ae-4767-9a96-f309bc41ddf3" Dec 10 18:57:12 crc kubenswrapper[4828]: I1210 18:57:12.787891 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:57:12 crc kubenswrapper[4828]: I1210 18:57:12.788005 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:12 crc kubenswrapper[4828]: I1210 18:57:12.788132 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:12 crc kubenswrapper[4828]: I1210 18:57:12.788281 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:12 crc kubenswrapper[4828]: E1210 18:57:12.788277 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:57:12 crc kubenswrapper[4828]: E1210 18:57:12.788423 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:12 crc kubenswrapper[4828]: E1210 18:57:12.788530 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:12 crc kubenswrapper[4828]: E1210 18:57:12.788609 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:13 crc kubenswrapper[4828]: I1210 18:57:13.372510 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xdxgd_c6e55203-65ae-4767-9a96-f309bc41ddf3/kube-multus/1.log" Dec 10 18:57:14 crc kubenswrapper[4828]: I1210 18:57:14.788496 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:14 crc kubenswrapper[4828]: I1210 18:57:14.788555 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:57:14 crc kubenswrapper[4828]: I1210 18:57:14.788583 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:14 crc kubenswrapper[4828]: E1210 18:57:14.789939 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:57:14 crc kubenswrapper[4828]: E1210 18:57:14.789658 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:14 crc kubenswrapper[4828]: E1210 18:57:14.790009 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:14 crc kubenswrapper[4828]: I1210 18:57:14.788736 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:14 crc kubenswrapper[4828]: E1210 18:57:14.790117 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:16 crc kubenswrapper[4828]: I1210 18:57:16.787712 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:16 crc kubenswrapper[4828]: I1210 18:57:16.787764 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:16 crc kubenswrapper[4828]: I1210 18:57:16.787712 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:16 crc kubenswrapper[4828]: E1210 18:57:16.787929 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:16 crc kubenswrapper[4828]: E1210 18:57:16.788107 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:16 crc kubenswrapper[4828]: E1210 18:57:16.788188 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:16 crc kubenswrapper[4828]: I1210 18:57:16.787737 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:57:16 crc kubenswrapper[4828]: E1210 18:57:16.788590 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:57:18 crc kubenswrapper[4828]: I1210 18:57:18.788236 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:18 crc kubenswrapper[4828]: I1210 18:57:18.788234 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:18 crc kubenswrapper[4828]: I1210 18:57:18.788318 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:57:18 crc kubenswrapper[4828]: I1210 18:57:18.788454 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:18 crc kubenswrapper[4828]: E1210 18:57:18.788649 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:18 crc kubenswrapper[4828]: E1210 18:57:18.788876 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:18 crc kubenswrapper[4828]: E1210 18:57:18.789249 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:57:18 crc kubenswrapper[4828]: E1210 18:57:18.789499 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:19 crc kubenswrapper[4828]: E1210 18:57:19.818369 4828 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 10 18:57:19 crc kubenswrapper[4828]: E1210 18:57:19.869374 4828 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 18:57:20 crc kubenswrapper[4828]: I1210 18:57:20.787720 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:20 crc kubenswrapper[4828]: I1210 18:57:20.787837 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:57:20 crc kubenswrapper[4828]: I1210 18:57:20.787792 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:20 crc kubenswrapper[4828]: I1210 18:57:20.787727 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:20 crc kubenswrapper[4828]: E1210 18:57:20.788002 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:20 crc kubenswrapper[4828]: E1210 18:57:20.788151 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:57:20 crc kubenswrapper[4828]: E1210 18:57:20.788221 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:20 crc kubenswrapper[4828]: E1210 18:57:20.788302 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:22 crc kubenswrapper[4828]: I1210 18:57:22.788111 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:22 crc kubenswrapper[4828]: I1210 18:57:22.788159 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:22 crc kubenswrapper[4828]: I1210 18:57:22.788194 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:57:22 crc kubenswrapper[4828]: E1210 18:57:22.788263 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:22 crc kubenswrapper[4828]: I1210 18:57:22.788293 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:22 crc kubenswrapper[4828]: E1210 18:57:22.788451 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:57:22 crc kubenswrapper[4828]: E1210 18:57:22.788650 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:22 crc kubenswrapper[4828]: E1210 18:57:22.789171 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:22 crc kubenswrapper[4828]: I1210 18:57:22.789457 4828 scope.go:117] "RemoveContainer" containerID="7e0ef04a88803858aabdbc8c1b702a19cc044e5df4e33e4d1a719f4d506e7c82" Dec 10 18:57:23 crc kubenswrapper[4828]: I1210 18:57:23.406625 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovnkube-controller/3.log" Dec 10 18:57:23 crc kubenswrapper[4828]: I1210 18:57:23.409213 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerStarted","Data":"a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867"} Dec 10 18:57:23 crc kubenswrapper[4828]: I1210 18:57:23.409629 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:57:23 crc kubenswrapper[4828]: I1210 18:57:23.437637 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podStartSLOduration=105.43760425400001 podStartE2EDuration="1m45.437604254s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:23.434516867 +0000 UTC m=+123.945127882" watchObservedRunningTime="2025-12-10 18:57:23.437604254 +0000 UTC m=+123.948215279" Dec 10 18:57:23 crc kubenswrapper[4828]: I1210 18:57:23.717503 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-74lmx"] Dec 10 18:57:23 crc kubenswrapper[4828]: I1210 18:57:23.717681 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:57:23 crc kubenswrapper[4828]: E1210 18:57:23.717866 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:57:24 crc kubenswrapper[4828]: I1210 18:57:24.788538 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:24 crc kubenswrapper[4828]: I1210 18:57:24.788640 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:24 crc kubenswrapper[4828]: E1210 18:57:24.788698 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:24 crc kubenswrapper[4828]: I1210 18:57:24.788640 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:24 crc kubenswrapper[4828]: E1210 18:57:24.788788 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:24 crc kubenswrapper[4828]: E1210 18:57:24.788990 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:24 crc kubenswrapper[4828]: E1210 18:57:24.871114 4828 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 18:57:25 crc kubenswrapper[4828]: I1210 18:57:25.788851 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:57:25 crc kubenswrapper[4828]: E1210 18:57:25.789112 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:57:26 crc kubenswrapper[4828]: I1210 18:57:26.788154 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:26 crc kubenswrapper[4828]: I1210 18:57:26.788345 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:26 crc kubenswrapper[4828]: I1210 18:57:26.788430 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:26 crc kubenswrapper[4828]: E1210 18:57:26.788519 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:26 crc kubenswrapper[4828]: E1210 18:57:26.788355 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:26 crc kubenswrapper[4828]: E1210 18:57:26.788604 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:27 crc kubenswrapper[4828]: I1210 18:57:27.788280 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:57:27 crc kubenswrapper[4828]: E1210 18:57:27.788722 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:57:27 crc kubenswrapper[4828]: I1210 18:57:27.789449 4828 scope.go:117] "RemoveContainer" containerID="bbe415208b7c4a043d3a9466d81931a16b2d600a3b5605eb89c7d2c2ffb1b71b" Dec 10 18:57:28 crc kubenswrapper[4828]: I1210 18:57:28.431186 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xdxgd_c6e55203-65ae-4767-9a96-f309bc41ddf3/kube-multus/1.log" Dec 10 18:57:28 crc kubenswrapper[4828]: I1210 18:57:28.431638 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xdxgd" event={"ID":"c6e55203-65ae-4767-9a96-f309bc41ddf3","Type":"ContainerStarted","Data":"ea927775625b0de792d093bd00de1d34b82f3826d19fa5ddaf39be9b182f20ab"} Dec 10 18:57:28 crc kubenswrapper[4828]: I1210 18:57:28.788388 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:28 crc kubenswrapper[4828]: I1210 18:57:28.788412 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:28 crc kubenswrapper[4828]: E1210 18:57:28.788539 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 10 18:57:28 crc kubenswrapper[4828]: E1210 18:57:28.788638 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 10 18:57:28 crc kubenswrapper[4828]: I1210 18:57:28.789234 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:28 crc kubenswrapper[4828]: E1210 18:57:28.789517 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 10 18:57:29 crc kubenswrapper[4828]: I1210 18:57:29.788146 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:57:29 crc kubenswrapper[4828]: E1210 18:57:29.791203 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-74lmx" podUID="472ff1d2-0936-4ee0-b738-dc26e2c03559" Dec 10 18:57:30 crc kubenswrapper[4828]: I1210 18:57:30.788678 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:30 crc kubenswrapper[4828]: I1210 18:57:30.788683 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:30 crc kubenswrapper[4828]: I1210 18:57:30.788708 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:30 crc kubenswrapper[4828]: I1210 18:57:30.791453 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 10 18:57:30 crc kubenswrapper[4828]: I1210 18:57:30.791516 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 10 18:57:30 crc kubenswrapper[4828]: I1210 18:57:30.792705 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 10 18:57:30 crc kubenswrapper[4828]: I1210 18:57:30.793611 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 10 18:57:31 crc kubenswrapper[4828]: I1210 18:57:31.787882 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:57:31 crc kubenswrapper[4828]: I1210 18:57:31.792130 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 10 18:57:31 crc kubenswrapper[4828]: I1210 18:57:31.792919 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.128275 4828 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.187094 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.187900 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.189735 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-flkll"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.190646 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.194871 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.196095 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.197365 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gspnq"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.197914 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.203324 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.203711 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.203754 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.204104 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.204219 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.204295 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.204392 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.204914 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.204938 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.206007 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2kkp7"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.214055 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.215826 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.234084 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.235362 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f92kp"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.235714 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.237120 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-m4fkj"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.237338 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f92kp" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.237985 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.238097 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hgsxd"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.238292 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.238444 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.238496 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n9pnl"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.238571 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.238931 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.239268 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.239609 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.241508 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.241677 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.241853 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.242074 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.242957 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.243265 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.244737 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.245130 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.245509 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/477e7e82-6ea9-4b53-8aa6-95507dbcc11c-service-ca-bundle\") pod \"authentication-operator-69f744f599-2kkp7\" (UID: \"477e7e82-6ea9-4b53-8aa6-95507dbcc11c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.245556 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxkql\" (UniqueName: \"kubernetes.io/projected/18526255-d66e-4a2d-85a0-433ce7c59757-kube-api-access-lxkql\") pod \"machine-approver-56656f9798-flkll\" (UID: \"18526255-d66e-4a2d-85a0-433ce7c59757\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.245624 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d9df65-e6cc-404d-8cf9-f50c566d5595-serving-cert\") pod \"route-controller-manager-6576b87f9c-j8vl2\" (UID: \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.245657 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/477e7e82-6ea9-4b53-8aa6-95507dbcc11c-serving-cert\") pod \"authentication-operator-69f744f599-2kkp7\" (UID: \"477e7e82-6ea9-4b53-8aa6-95507dbcc11c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.245692 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/de5e17c4-b5df-48a1-898c-436e2c0923c3-images\") pod \"machine-api-operator-5694c8668f-gspnq\" (UID: \"de5e17c4-b5df-48a1-898c-436e2c0923c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.245714 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/de5e17c4-b5df-48a1-898c-436e2c0923c3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gspnq\" (UID: \"de5e17c4-b5df-48a1-898c-436e2c0923c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.245751 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/18526255-d66e-4a2d-85a0-433ce7c59757-auth-proxy-config\") pod \"machine-approver-56656f9798-flkll\" (UID: \"18526255-d66e-4a2d-85a0-433ce7c59757\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.245773 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d9df65-e6cc-404d-8cf9-f50c566d5595-config\") pod \"route-controller-manager-6576b87f9c-j8vl2\" (UID: \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.245812 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18526255-d66e-4a2d-85a0-433ce7c59757-config\") pod \"machine-approver-56656f9798-flkll\" (UID: \"18526255-d66e-4a2d-85a0-433ce7c59757\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.245836 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/18526255-d66e-4a2d-85a0-433ce7c59757-machine-approver-tls\") pod \"machine-approver-56656f9798-flkll\" (UID: \"18526255-d66e-4a2d-85a0-433ce7c59757\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.245861 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477e7e82-6ea9-4b53-8aa6-95507dbcc11c-config\") pod \"authentication-operator-69f744f599-2kkp7\" (UID: \"477e7e82-6ea9-4b53-8aa6-95507dbcc11c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.245880 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/477e7e82-6ea9-4b53-8aa6-95507dbcc11c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2kkp7\" (UID: \"477e7e82-6ea9-4b53-8aa6-95507dbcc11c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.245908 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9d9df65-e6cc-404d-8cf9-f50c566d5595-client-ca\") pod \"route-controller-manager-6576b87f9c-j8vl2\" (UID: \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.245927 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de5e17c4-b5df-48a1-898c-436e2c0923c3-config\") pod \"machine-api-operator-5694c8668f-gspnq\" (UID: \"de5e17c4-b5df-48a1-898c-436e2c0923c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.245939 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.246708 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.245945 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph6xt\" (UniqueName: \"kubernetes.io/projected/de5e17c4-b5df-48a1-898c-436e2c0923c3-kube-api-access-ph6xt\") pod \"machine-api-operator-5694c8668f-gspnq\" (UID: \"de5e17c4-b5df-48a1-898c-436e2c0923c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.247038 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nwbk\" (UniqueName: \"kubernetes.io/projected/e9d9df65-e6cc-404d-8cf9-f50c566d5595-kube-api-access-4nwbk\") pod \"route-controller-manager-6576b87f9c-j8vl2\" (UID: \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.247106 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfz82\" (UniqueName: \"kubernetes.io/projected/477e7e82-6ea9-4b53-8aa6-95507dbcc11c-kube-api-access-bfz82\") pod \"authentication-operator-69f744f599-2kkp7\" (UID: \"477e7e82-6ea9-4b53-8aa6-95507dbcc11c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.248901 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.249392 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.250295 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mhrd"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.250997 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mhrd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.253839 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.255098 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.255363 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-25pdm"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.257712 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-s4zkd"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.258263 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-s4zkd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.258548 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-25pdm" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.260527 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.261607 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.261853 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.262093 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.262362 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.262451 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.262551 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.263709 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8xs5l"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.265063 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.265582 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.265858 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c4wpt"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.266250 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.270907 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.271142 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-x4r6k"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.272174 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-x4r6k" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.272456 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.272579 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.274030 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vrmnl"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.275269 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vrmnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.276045 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vfsz7"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.276455 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.278886 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7vh2m"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.279580 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7vh2m" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.286759 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.286904 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.287161 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.287328 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.288030 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.288329 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.288402 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.288564 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.288840 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.289275 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.291260 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.292640 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.292723 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.292977 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.293317 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.293431 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.293654 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.293759 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.293881 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.294029 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.294187 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.294308 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.294394 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.294746 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.294930 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.295539 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.295731 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.297132 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.299520 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.299618 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.299902 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.300028 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.300183 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.300346 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.301005 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.301022 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.301179 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.301206 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.301207 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.309927 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.310091 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.310217 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.310632 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.310814 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.310917 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.311543 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.309927 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.310092 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.312022 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.312142 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.312244 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.314526 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.314695 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.314847 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.314870 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-qtdmw"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.315079 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.315164 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.315392 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gdt2z"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.315521 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.315631 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.315875 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gdt2z" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.317551 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.318916 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncvkn"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.319566 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncvkn" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.324406 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.324688 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.324855 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.324972 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.325156 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.329066 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nscjn"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.330102 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nscjn" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.330370 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mlzfg"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.333940 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.334772 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mlzfg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.337387 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.342002 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w7n9w"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.342747 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.346085 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.346312 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mqr27"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347029 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-mqr27" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347524 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e389c321-1fee-4503-8f27-ce2eb2dd76c3-encryption-config\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347552 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d97e08b3-79e2-434a-8de9-486cac9a7f19-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7mhrd\" (UID: \"d97e08b3-79e2-434a-8de9-486cac9a7f19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mhrd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347571 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/de8e8906-1b66-4e41-8881-513f911a7407-etcd-ca\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347587 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k66vf\" (UniqueName: \"kubernetes.io/projected/109ecad0-ad50-4121-843e-319c8638705f-kube-api-access-k66vf\") pod \"migrator-59844c95c7-gdt2z\" (UID: \"109ecad0-ad50-4121-843e-319c8638705f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gdt2z" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347605 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09f4f7a0-5033-482d-9333-a2255ae9fb8c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncvkn\" (UID: \"09f4f7a0-5033-482d-9333-a2255ae9fb8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncvkn" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347620 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f59b0400-f54d-47cc-a58f-7a0399fa857d-audit-policies\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347633 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e621a558-87ca-499b-8d99-a3584988ece6-trusted-ca\") pod \"console-operator-58897d9998-25pdm\" (UID: \"e621a558-87ca-499b-8d99-a3584988ece6\") " pod="openshift-console-operator/console-operator-58897d9998-25pdm" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347649 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2b3764a2-69cd-4f7a-ac6e-5250925117e9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mlzfg\" (UID: \"2b3764a2-69cd-4f7a-ac6e-5250925117e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mlzfg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347669 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d9df65-e6cc-404d-8cf9-f50c566d5595-serving-cert\") pod \"route-controller-manager-6576b87f9c-j8vl2\" (UID: \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347684 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/2cf82e3e-7ba7-4a3b-adf5-c635561b63d3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-x4r6k\" (UID: \"2cf82e3e-7ba7-4a3b-adf5-c635561b63d3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-x4r6k" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347699 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2295ac8e-aa01-4349-a4a3-843e45df1475-config\") pod \"kube-apiserver-operator-766d6c64bb-7vh2m\" (UID: \"2295ac8e-aa01-4349-a4a3-843e45df1475\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7vh2m" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347715 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-console-config\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347731 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b3764a2-69cd-4f7a-ac6e-5250925117e9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mlzfg\" (UID: \"2b3764a2-69cd-4f7a-ac6e-5250925117e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mlzfg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347746 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e621a558-87ca-499b-8d99-a3584988ece6-config\") pod \"console-operator-58897d9998-25pdm\" (UID: \"e621a558-87ca-499b-8d99-a3584988ece6\") " pod="openshift-console-operator/console-operator-58897d9998-25pdm" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347760 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjxw9\" (UniqueName: \"kubernetes.io/projected/de8e8906-1b66-4e41-8881-513f911a7407-kube-api-access-fjxw9\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347775 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/53029d0b-3ada-4e72-856c-381ab7ae6292-audit-dir\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347815 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1ca9c41a-969e-4475-8487-78f6f4d8e6b8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vfk86\" (UID: \"1ca9c41a-969e-4475-8487-78f6f4d8e6b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347823 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347975 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.351352 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.347832 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.351557 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99680c81-94dd-4f90-9133-1e572dd9f34c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-f92kp\" (UID: \"99680c81-94dd-4f90-9133-1e572dd9f34c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f92kp" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.351625 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/18526255-d66e-4a2d-85a0-433ce7c59757-auth-proxy-config\") pod \"machine-approver-56656f9798-flkll\" (UID: \"18526255-d66e-4a2d-85a0-433ce7c59757\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.351668 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f59b0400-f54d-47cc-a58f-7a0399fa857d-encryption-config\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.351709 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d9df65-e6cc-404d-8cf9-f50c566d5595-config\") pod \"route-controller-manager-6576b87f9c-j8vl2\" (UID: \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.351742 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d9076471-a7f4-437b-82fc-5384b61cc048-proxy-tls\") pod \"machine-config-operator-74547568cd-mm2tg\" (UID: \"d9076471-a7f4-437b-82fc-5384b61cc048\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.351781 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e389c321-1fee-4503-8f27-ce2eb2dd76c3-audit-dir\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.351835 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j2cc\" (UniqueName: \"kubernetes.io/projected/2cf82e3e-7ba7-4a3b-adf5-c635561b63d3-kube-api-access-6j2cc\") pod \"cluster-samples-operator-665b6dd947-x4r6k\" (UID: \"2cf82e3e-7ba7-4a3b-adf5-c635561b63d3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-x4r6k" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.351921 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkj69\" (UniqueName: \"kubernetes.io/projected/d97e08b3-79e2-434a-8de9-486cac9a7f19-kube-api-access-nkj69\") pod \"openshift-apiserver-operator-796bbdcf4f-7mhrd\" (UID: \"d97e08b3-79e2-434a-8de9-486cac9a7f19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mhrd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.351949 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a2629f52-3b35-4707-8a24-53d7bbb70465-console-oauth-config\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.351984 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjc88\" (UniqueName: \"kubernetes.io/projected/6bae8a54-7dad-448f-b513-0726b7dbce91-kube-api-access-zjc88\") pod \"cluster-image-registry-operator-dc59b4c8b-rhm46\" (UID: \"6bae8a54-7dad-448f-b513-0726b7dbce91\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.352018 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d9076471-a7f4-437b-82fc-5384b61cc048-images\") pod \"machine-config-operator-74547568cd-mm2tg\" (UID: \"d9076471-a7f4-437b-82fc-5384b61cc048\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.352034 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.352058 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/477e7e82-6ea9-4b53-8aa6-95507dbcc11c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2kkp7\" (UID: \"477e7e82-6ea9-4b53-8aa6-95507dbcc11c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.352090 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc6kx\" (UniqueName: \"kubernetes.io/projected/62e1b21d-36c2-4047-8913-8fc45e0401d7-kube-api-access-tc6kx\") pod \"dns-operator-744455d44c-vrmnl\" (UID: \"62e1b21d-36c2-4047-8913-8fc45e0401d7\") " pod="openshift-dns-operator/dns-operator-744455d44c-vrmnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.352119 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f59b0400-f54d-47cc-a58f-7a0399fa857d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.352324 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/18526255-d66e-4a2d-85a0-433ce7c59757-auth-proxy-config\") pod \"machine-approver-56656f9798-flkll\" (UID: \"18526255-d66e-4a2d-85a0-433ce7c59757\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.352346 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/09f4f7a0-5033-482d-9333-a2255ae9fb8c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncvkn\" (UID: \"09f4f7a0-5033-482d-9333-a2255ae9fb8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncvkn" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.352389 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9592c228-7f94-4923-818e-2d307bdf8d28-serving-cert\") pod \"openshift-config-operator-7777fb866f-zrqbh\" (UID: \"9592c228-7f94-4923-818e-2d307bdf8d28\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.352478 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9d9df65-e6cc-404d-8cf9-f50c566d5595-client-ca\") pod \"route-controller-manager-6576b87f9c-j8vl2\" (UID: \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.352645 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nwbk\" (UniqueName: \"kubernetes.io/projected/e9d9df65-e6cc-404d-8cf9-f50c566d5595-kube-api-access-4nwbk\") pod \"route-controller-manager-6576b87f9c-j8vl2\" (UID: \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.352676 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de5e17c4-b5df-48a1-898c-436e2c0923c3-config\") pod \"machine-api-operator-5694c8668f-gspnq\" (UID: \"de5e17c4-b5df-48a1-898c-436e2c0923c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.352699 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9bl4\" (UniqueName: \"kubernetes.io/projected/53029d0b-3ada-4e72-856c-381ab7ae6292-kube-api-access-k9bl4\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.353311 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d9df65-e6cc-404d-8cf9-f50c566d5595-config\") pod \"route-controller-manager-6576b87f9c-j8vl2\" (UID: \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.353663 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.353832 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9a8d7330-30c8-4a9c-ae2f-695b2c17deea-bound-sa-token\") pod \"ingress-operator-5b745b69d9-p46hd\" (UID: \"9a8d7330-30c8-4a9c-ae2f-695b2c17deea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.353891 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2295ac8e-aa01-4349-a4a3-843e45df1475-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7vh2m\" (UID: \"2295ac8e-aa01-4349-a4a3-843e45df1475\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7vh2m" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354096 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/477e7e82-6ea9-4b53-8aa6-95507dbcc11c-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2kkp7\" (UID: \"477e7e82-6ea9-4b53-8aa6-95507dbcc11c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354129 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bq986\" (UniqueName: \"kubernetes.io/projected/d9076471-a7f4-437b-82fc-5384b61cc048-kube-api-access-bq986\") pod \"machine-config-operator-74547568cd-mm2tg\" (UID: \"d9076471-a7f4-437b-82fc-5384b61cc048\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354168 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/47cfb7d6-3723-4269-95ec-f6a512335387-stats-auth\") pod \"router-default-5444994796-qtdmw\" (UID: \"47cfb7d6-3723-4269-95ec-f6a512335387\") " pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354198 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-audit-policies\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354201 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9d9df65-e6cc-404d-8cf9-f50c566d5595-client-ca\") pod \"route-controller-manager-6576b87f9c-j8vl2\" (UID: \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354229 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a2629f52-3b35-4707-8a24-53d7bbb70465-console-serving-cert\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354274 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9a8d7330-30c8-4a9c-ae2f-695b2c17deea-metrics-tls\") pod \"ingress-operator-5b745b69d9-p46hd\" (UID: \"9a8d7330-30c8-4a9c-ae2f-695b2c17deea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354281 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7x5g"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354315 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e389c321-1fee-4503-8f27-ce2eb2dd76c3-audit\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354342 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6bae8a54-7dad-448f-b513-0726b7dbce91-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rhm46\" (UID: \"6bae8a54-7dad-448f-b513-0726b7dbce91\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354385 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfz82\" (UniqueName: \"kubernetes.io/projected/477e7e82-6ea9-4b53-8aa6-95507dbcc11c-kube-api-access-bfz82\") pod \"authentication-operator-69f744f599-2kkp7\" (UID: \"477e7e82-6ea9-4b53-8aa6-95507dbcc11c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354429 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354456 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354523 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09f4f7a0-5033-482d-9333-a2255ae9fb8c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncvkn\" (UID: \"09f4f7a0-5033-482d-9333-a2255ae9fb8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncvkn" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354541 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e621a558-87ca-499b-8d99-a3584988ece6-serving-cert\") pod \"console-operator-58897d9998-25pdm\" (UID: \"e621a558-87ca-499b-8d99-a3584988ece6\") " pod="openshift-console-operator/console-operator-58897d9998-25pdm" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354618 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354640 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/de8e8906-1b66-4e41-8881-513f911a7407-etcd-client\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.354684 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f59b0400-f54d-47cc-a58f-7a0399fa857d-etcd-client\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.355125 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f59b0400-f54d-47cc-a58f-7a0399fa857d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.355180 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.355210 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99680c81-94dd-4f90-9133-1e572dd9f34c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-f92kp\" (UID: \"99680c81-94dd-4f90-9133-1e572dd9f34c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f92kp" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.355254 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/477e7e82-6ea9-4b53-8aa6-95507dbcc11c-service-ca-bundle\") pod \"authentication-operator-69f744f599-2kkp7\" (UID: \"477e7e82-6ea9-4b53-8aa6-95507dbcc11c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.355279 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxkql\" (UniqueName: \"kubernetes.io/projected/18526255-d66e-4a2d-85a0-433ce7c59757-kube-api-access-lxkql\") pod \"machine-approver-56656f9798-flkll\" (UID: \"18526255-d66e-4a2d-85a0-433ce7c59757\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.355310 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-trusted-ca-bundle\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.355746 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7x5g" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.355924 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e389c321-1fee-4503-8f27-ce2eb2dd76c3-config\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.355974 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.356004 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9a8d7330-30c8-4a9c-ae2f-695b2c17deea-trusted-ca\") pod \"ingress-operator-5b745b69d9-p46hd\" (UID: \"9a8d7330-30c8-4a9c-ae2f-695b2c17deea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.356322 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e389c321-1fee-4503-8f27-ce2eb2dd76c3-image-import-ca\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.356374 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bngnt\" (UniqueName: \"kubernetes.io/projected/510cc49c-3ed7-4998-9d29-c50a7ab86337-kube-api-access-bngnt\") pod \"downloads-7954f5f757-s4zkd\" (UID: \"510cc49c-3ed7-4998-9d29-c50a7ab86337\") " pod="openshift-console/downloads-7954f5f757-s4zkd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.356413 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/477e7e82-6ea9-4b53-8aa6-95507dbcc11c-serving-cert\") pod \"authentication-operator-69f744f599-2kkp7\" (UID: \"477e7e82-6ea9-4b53-8aa6-95507dbcc11c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.356449 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47cfb7d6-3723-4269-95ec-f6a512335387-service-ca-bundle\") pod \"router-default-5444994796-qtdmw\" (UID: \"47cfb7d6-3723-4269-95ec-f6a512335387\") " pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.356485 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2295ac8e-aa01-4349-a4a3-843e45df1475-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7vh2m\" (UID: \"2295ac8e-aa01-4349-a4a3-843e45df1475\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7vh2m" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.356771 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357013 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzflq\" (UniqueName: \"kubernetes.io/projected/e621a558-87ca-499b-8d99-a3584988ece6-kube-api-access-xzflq\") pod \"console-operator-58897d9998-25pdm\" (UID: \"e621a558-87ca-499b-8d99-a3584988ece6\") " pod="openshift-console-operator/console-operator-58897d9998-25pdm" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357066 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d97e08b3-79e2-434a-8de9-486cac9a7f19-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7mhrd\" (UID: \"d97e08b3-79e2-434a-8de9-486cac9a7f19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mhrd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357104 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357118 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357139 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-client-ca\") pod \"controller-manager-879f6c89f-n9pnl\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357378 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/de5e17c4-b5df-48a1-898c-436e2c0923c3-images\") pod \"machine-api-operator-5694c8668f-gspnq\" (UID: \"de5e17c4-b5df-48a1-898c-436e2c0923c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357485 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/de5e17c4-b5df-48a1-898c-436e2c0923c3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gspnq\" (UID: \"de5e17c4-b5df-48a1-898c-436e2c0923c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357518 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e389c321-1fee-4503-8f27-ce2eb2dd76c3-etcd-serving-ca\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357543 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq5cn\" (UniqueName: \"kubernetes.io/projected/5f498d61-1987-48a1-82d2-14ea6db95c84-kube-api-access-fq5cn\") pod \"controller-manager-879f6c89f-n9pnl\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357589 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgvzq\" (UniqueName: \"kubernetes.io/projected/a2629f52-3b35-4707-8a24-53d7bbb70465-kube-api-access-vgvzq\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357612 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9592c228-7f94-4923-818e-2d307bdf8d28-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zrqbh\" (UID: \"9592c228-7f94-4923-818e-2d307bdf8d28\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357637 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snj5z\" (UniqueName: \"kubernetes.io/projected/9a8d7330-30c8-4a9c-ae2f-695b2c17deea-kube-api-access-snj5z\") pod \"ingress-operator-5b745b69d9-p46hd\" (UID: \"9a8d7330-30c8-4a9c-ae2f-695b2c17deea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357661 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de5e17c4-b5df-48a1-898c-436e2c0923c3-config\") pod \"machine-api-operator-5694c8668f-gspnq\" (UID: \"de5e17c4-b5df-48a1-898c-436e2c0923c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357658 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs4d5\" (UniqueName: \"kubernetes.io/projected/9592c228-7f94-4923-818e-2d307bdf8d28-kube-api-access-rs4d5\") pod \"openshift-config-operator-7777fb866f-zrqbh\" (UID: \"9592c228-7f94-4923-818e-2d307bdf8d28\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357691 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357723 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/477e7e82-6ea9-4b53-8aa6-95507dbcc11c-service-ca-bundle\") pod \"authentication-operator-69f744f599-2kkp7\" (UID: \"477e7e82-6ea9-4b53-8aa6-95507dbcc11c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357891 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357953 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nltk\" (UniqueName: \"kubernetes.io/projected/99680c81-94dd-4f90-9133-1e572dd9f34c-kube-api-access-6nltk\") pod \"openshift-controller-manager-operator-756b6f6bc6-f92kp\" (UID: \"99680c81-94dd-4f90-9133-1e572dd9f34c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f92kp" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357974 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e389c321-1fee-4503-8f27-ce2eb2dd76c3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.357994 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de8e8906-1b66-4e41-8881-513f911a7407-serving-cert\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.358036 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llcd7\" (UniqueName: \"kubernetes.io/projected/f59b0400-f54d-47cc-a58f-7a0399fa857d-kube-api-access-llcd7\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.358058 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-config\") pod \"controller-manager-879f6c89f-n9pnl\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.358075 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzz6l\" (UniqueName: \"kubernetes.io/projected/e389c321-1fee-4503-8f27-ce2eb2dd76c3-kube-api-access-zzz6l\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.358226 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18526255-d66e-4a2d-85a0-433ce7c59757-config\") pod \"machine-approver-56656f9798-flkll\" (UID: \"18526255-d66e-4a2d-85a0-433ce7c59757\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.358273 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f59b0400-f54d-47cc-a58f-7a0399fa857d-serving-cert\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.358308 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/18526255-d66e-4a2d-85a0-433ce7c59757-machine-approver-tls\") pod \"machine-approver-56656f9798-flkll\" (UID: \"18526255-d66e-4a2d-85a0-433ce7c59757\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.358341 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b3764a2-69cd-4f7a-ac6e-5250925117e9-config\") pod \"kube-controller-manager-operator-78b949d7b-mlzfg\" (UID: \"2b3764a2-69cd-4f7a-ac6e-5250925117e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mlzfg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.358922 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f68aa373-3441-49f3-abc0-d5647d7bd869-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-nscjn\" (UID: \"f68aa373-3441-49f3-abc0-d5647d7bd869\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nscjn" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.358966 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/6bae8a54-7dad-448f-b513-0726b7dbce91-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rhm46\" (UID: \"6bae8a54-7dad-448f-b513-0726b7dbce91\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.359007 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477e7e82-6ea9-4b53-8aa6-95507dbcc11c-config\") pod \"authentication-operator-69f744f599-2kkp7\" (UID: \"477e7e82-6ea9-4b53-8aa6-95507dbcc11c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.359039 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-oauth-serving-cert\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.359119 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/de5e17c4-b5df-48a1-898c-436e2c0923c3-images\") pod \"machine-api-operator-5694c8668f-gspnq\" (UID: \"de5e17c4-b5df-48a1-898c-436e2c0923c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.359210 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e389c321-1fee-4503-8f27-ce2eb2dd76c3-node-pullsecrets\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.360034 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1ca9c41a-969e-4475-8487-78f6f4d8e6b8-proxy-tls\") pod \"machine-config-controller-84d6567774-vfk86\" (UID: \"1ca9c41a-969e-4475-8487-78f6f4d8e6b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.360070 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/de8e8906-1b66-4e41-8881-513f911a7407-etcd-service-ca\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.360092 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e389c321-1fee-4503-8f27-ce2eb2dd76c3-serving-cert\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.360121 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph6xt\" (UniqueName: \"kubernetes.io/projected/de5e17c4-b5df-48a1-898c-436e2c0923c3-kube-api-access-ph6xt\") pod \"machine-api-operator-5694c8668f-gspnq\" (UID: \"de5e17c4-b5df-48a1-898c-436e2c0923c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.360142 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-n9pnl\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.360163 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47cfb7d6-3723-4269-95ec-f6a512335387-metrics-certs\") pod \"router-default-5444994796-qtdmw\" (UID: \"47cfb7d6-3723-4269-95ec-f6a512335387\") " pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.360181 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.360205 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f498d61-1987-48a1-82d2-14ea6db95c84-serving-cert\") pod \"controller-manager-879f6c89f-n9pnl\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.360225 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/62e1b21d-36c2-4047-8913-8fc45e0401d7-metrics-tls\") pod \"dns-operator-744455d44c-vrmnl\" (UID: \"62e1b21d-36c2-4047-8913-8fc45e0401d7\") " pod="openshift-dns-operator/dns-operator-744455d44c-vrmnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.360244 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d9076471-a7f4-437b-82fc-5384b61cc048-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mm2tg\" (UID: \"d9076471-a7f4-437b-82fc-5384b61cc048\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.360325 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.360348 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bghh9\" (UniqueName: \"kubernetes.io/projected/1ca9c41a-969e-4475-8487-78f6f4d8e6b8-kube-api-access-bghh9\") pod \"machine-config-controller-84d6567774-vfk86\" (UID: \"1ca9c41a-969e-4475-8487-78f6f4d8e6b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.360380 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-service-ca\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.360402 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pf6z\" (UniqueName: \"kubernetes.io/projected/f68aa373-3441-49f3-abc0-d5647d7bd869-kube-api-access-4pf6z\") pod \"control-plane-machine-set-operator-78cbb6b69f-nscjn\" (UID: \"f68aa373-3441-49f3-abc0-d5647d7bd869\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nscjn" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.360421 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72b2f\" (UniqueName: \"kubernetes.io/projected/47cfb7d6-3723-4269-95ec-f6a512335387-kube-api-access-72b2f\") pod \"router-default-5444994796-qtdmw\" (UID: \"47cfb7d6-3723-4269-95ec-f6a512335387\") " pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.360443 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f59b0400-f54d-47cc-a58f-7a0399fa857d-audit-dir\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.360467 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/47cfb7d6-3723-4269-95ec-f6a512335387-default-certificate\") pod \"router-default-5444994796-qtdmw\" (UID: \"47cfb7d6-3723-4269-95ec-f6a512335387\") " pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.361010 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18526255-d66e-4a2d-85a0-433ce7c59757-config\") pod \"machine-approver-56656f9798-flkll\" (UID: \"18526255-d66e-4a2d-85a0-433ce7c59757\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.361685 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.362220 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e389c321-1fee-4503-8f27-ce2eb2dd76c3-etcd-client\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.362284 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6bae8a54-7dad-448f-b513-0726b7dbce91-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rhm46\" (UID: \"6bae8a54-7dad-448f-b513-0726b7dbce91\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.362310 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de8e8906-1b66-4e41-8881-513f911a7407-config\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.362967 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.364867 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/477e7e82-6ea9-4b53-8aa6-95507dbcc11c-config\") pod \"authentication-operator-69f744f599-2kkp7\" (UID: \"477e7e82-6ea9-4b53-8aa6-95507dbcc11c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.364923 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.368396 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/477e7e82-6ea9-4b53-8aa6-95507dbcc11c-serving-cert\") pod \"authentication-operator-69f744f599-2kkp7\" (UID: \"477e7e82-6ea9-4b53-8aa6-95507dbcc11c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.370224 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.370414 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.373836 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.373838 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/de5e17c4-b5df-48a1-898c-436e2c0923c3-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gspnq\" (UID: \"de5e17c4-b5df-48a1-898c-436e2c0923c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.374364 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.374536 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d9df65-e6cc-404d-8cf9-f50c566d5595-serving-cert\") pod \"route-controller-manager-6576b87f9c-j8vl2\" (UID: \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.375588 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.375928 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9jzvz"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.376537 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9jzvz" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.376629 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.376739 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.378013 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.378861 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wd865"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.379000 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/18526255-d66e-4a2d-85a0-433ce7c59757-machine-approver-tls\") pod \"machine-approver-56656f9798-flkll\" (UID: \"18526255-d66e-4a2d-85a0-433ce7c59757\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.380094 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2kkp7"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.380184 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wd865" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.381119 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.382028 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.382212 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gspnq"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.383177 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fv8kj"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.384981 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.385112 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fv8kj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.385442 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hgsxd"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.386495 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.387525 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-25pdm"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.388515 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n9pnl"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.390029 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8xs5l"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.390640 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.391691 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vfsz7"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.392645 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.392678 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nscjn"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.393671 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-hgg7d"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.394765 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mhrd"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.394875 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hgg7d" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.395822 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c4wpt"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.397890 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wd865"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.400690 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-x4r6k"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.400714 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w7n9w"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.401969 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-m4fkj"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.406850 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.409421 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncvkn"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.410255 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.410640 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.414255 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.415527 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-s4zkd"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.416889 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gdt2z"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.417902 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7vh2m"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.419232 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f92kp"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.421087 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mqr27"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.422346 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-22lss"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.423055 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-22lss" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.423376 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-bwnph"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.423747 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-bwnph" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.424423 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.425562 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mlzfg"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.426626 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fv8kj"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.427832 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-bwnph"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.428898 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7x5g"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.429921 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.429984 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vrmnl"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.430920 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9jzvz"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.431886 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.432787 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.436682 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hgg7d"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.443664 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-z2f4h"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.445387 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.445659 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-z2f4h"] Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.450355 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.465119 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9a8d7330-30c8-4a9c-ae2f-695b2c17deea-bound-sa-token\") pod \"ingress-operator-5b745b69d9-p46hd\" (UID: \"9a8d7330-30c8-4a9c-ae2f-695b2c17deea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.465151 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2295ac8e-aa01-4349-a4a3-843e45df1475-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7vh2m\" (UID: \"2295ac8e-aa01-4349-a4a3-843e45df1475\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7vh2m" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.465176 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/47cfb7d6-3723-4269-95ec-f6a512335387-stats-auth\") pod \"router-default-5444994796-qtdmw\" (UID: \"47cfb7d6-3723-4269-95ec-f6a512335387\") " pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.465195 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-audit-policies\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.465224 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bq986\" (UniqueName: \"kubernetes.io/projected/d9076471-a7f4-437b-82fc-5384b61cc048-kube-api-access-bq986\") pod \"machine-config-operator-74547568cd-mm2tg\" (UID: \"d9076471-a7f4-437b-82fc-5384b61cc048\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.465245 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e389c321-1fee-4503-8f27-ce2eb2dd76c3-audit\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.465262 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6bae8a54-7dad-448f-b513-0726b7dbce91-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rhm46\" (UID: \"6bae8a54-7dad-448f-b513-0726b7dbce91\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.465281 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a2629f52-3b35-4707-8a24-53d7bbb70465-console-serving-cert\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.465297 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9a8d7330-30c8-4a9c-ae2f-695b2c17deea-metrics-tls\") pod \"ingress-operator-5b745b69d9-p46hd\" (UID: \"9a8d7330-30c8-4a9c-ae2f-695b2c17deea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.465311 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.465335 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.465353 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09f4f7a0-5033-482d-9333-a2255ae9fb8c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncvkn\" (UID: \"09f4f7a0-5033-482d-9333-a2255ae9fb8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncvkn" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.465369 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e621a558-87ca-499b-8d99-a3584988ece6-serving-cert\") pod \"console-operator-58897d9998-25pdm\" (UID: \"e621a558-87ca-499b-8d99-a3584988ece6\") " pod="openshift-console-operator/console-operator-58897d9998-25pdm" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.465390 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.465526 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/de8e8906-1b66-4e41-8881-513f911a7407-etcd-client\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.465544 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f59b0400-f54d-47cc-a58f-7a0399fa857d-etcd-client\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466357 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466357 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e389c321-1fee-4503-8f27-ce2eb2dd76c3-audit\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466462 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466489 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99680c81-94dd-4f90-9133-1e572dd9f34c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-f92kp\" (UID: \"99680c81-94dd-4f90-9133-1e572dd9f34c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f92kp" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466523 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f59b0400-f54d-47cc-a58f-7a0399fa857d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466551 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-audit-policies\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466560 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-trusted-ca-bundle\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466661 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e389c321-1fee-4503-8f27-ce2eb2dd76c3-config\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466717 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466747 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9a8d7330-30c8-4a9c-ae2f-695b2c17deea-trusted-ca\") pod \"ingress-operator-5b745b69d9-p46hd\" (UID: \"9a8d7330-30c8-4a9c-ae2f-695b2c17deea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466751 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466768 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e389c321-1fee-4503-8f27-ce2eb2dd76c3-image-import-ca\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466817 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bngnt\" (UniqueName: \"kubernetes.io/projected/510cc49c-3ed7-4998-9d29-c50a7ab86337-kube-api-access-bngnt\") pod \"downloads-7954f5f757-s4zkd\" (UID: \"510cc49c-3ed7-4998-9d29-c50a7ab86337\") " pod="openshift-console/downloads-7954f5f757-s4zkd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466861 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47cfb7d6-3723-4269-95ec-f6a512335387-service-ca-bundle\") pod \"router-default-5444994796-qtdmw\" (UID: \"47cfb7d6-3723-4269-95ec-f6a512335387\") " pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466887 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2295ac8e-aa01-4349-a4a3-843e45df1475-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7vh2m\" (UID: \"2295ac8e-aa01-4349-a4a3-843e45df1475\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7vh2m" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466910 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzflq\" (UniqueName: \"kubernetes.io/projected/e621a558-87ca-499b-8d99-a3584988ece6-kube-api-access-xzflq\") pod \"console-operator-58897d9998-25pdm\" (UID: \"e621a558-87ca-499b-8d99-a3584988ece6\") " pod="openshift-console-operator/console-operator-58897d9998-25pdm" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466935 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d97e08b3-79e2-434a-8de9-486cac9a7f19-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7mhrd\" (UID: \"d97e08b3-79e2-434a-8de9-486cac9a7f19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mhrd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466960 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.466991 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-client-ca\") pod \"controller-manager-879f6c89f-n9pnl\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467011 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e389c321-1fee-4503-8f27-ce2eb2dd76c3-etcd-serving-ca\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467040 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq5cn\" (UniqueName: \"kubernetes.io/projected/5f498d61-1987-48a1-82d2-14ea6db95c84-kube-api-access-fq5cn\") pod \"controller-manager-879f6c89f-n9pnl\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467065 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgvzq\" (UniqueName: \"kubernetes.io/projected/a2629f52-3b35-4707-8a24-53d7bbb70465-kube-api-access-vgvzq\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467089 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9592c228-7f94-4923-818e-2d307bdf8d28-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zrqbh\" (UID: \"9592c228-7f94-4923-818e-2d307bdf8d28\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467114 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snj5z\" (UniqueName: \"kubernetes.io/projected/9a8d7330-30c8-4a9c-ae2f-695b2c17deea-kube-api-access-snj5z\") pod \"ingress-operator-5b745b69d9-p46hd\" (UID: \"9a8d7330-30c8-4a9c-ae2f-695b2c17deea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467133 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs4d5\" (UniqueName: \"kubernetes.io/projected/9592c228-7f94-4923-818e-2d307bdf8d28-kube-api-access-rs4d5\") pod \"openshift-config-operator-7777fb866f-zrqbh\" (UID: \"9592c228-7f94-4923-818e-2d307bdf8d28\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467139 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f59b0400-f54d-47cc-a58f-7a0399fa857d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467154 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467179 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nltk\" (UniqueName: \"kubernetes.io/projected/99680c81-94dd-4f90-9133-1e572dd9f34c-kube-api-access-6nltk\") pod \"openshift-controller-manager-operator-756b6f6bc6-f92kp\" (UID: \"99680c81-94dd-4f90-9133-1e572dd9f34c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f92kp" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467204 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llcd7\" (UniqueName: \"kubernetes.io/projected/f59b0400-f54d-47cc-a58f-7a0399fa857d-kube-api-access-llcd7\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467224 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e389c321-1fee-4503-8f27-ce2eb2dd76c3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467247 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e389c321-1fee-4503-8f27-ce2eb2dd76c3-config\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467249 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de8e8906-1b66-4e41-8881-513f911a7407-serving-cert\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467297 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-config\") pod \"controller-manager-879f6c89f-n9pnl\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467319 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzz6l\" (UniqueName: \"kubernetes.io/projected/e389c321-1fee-4503-8f27-ce2eb2dd76c3-kube-api-access-zzz6l\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467336 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f59b0400-f54d-47cc-a58f-7a0399fa857d-serving-cert\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467355 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b3764a2-69cd-4f7a-ac6e-5250925117e9-config\") pod \"kube-controller-manager-operator-78b949d7b-mlzfg\" (UID: \"2b3764a2-69cd-4f7a-ac6e-5250925117e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mlzfg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467380 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f68aa373-3441-49f3-abc0-d5647d7bd869-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-nscjn\" (UID: \"f68aa373-3441-49f3-abc0-d5647d7bd869\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nscjn" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467400 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/6bae8a54-7dad-448f-b513-0726b7dbce91-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rhm46\" (UID: \"6bae8a54-7dad-448f-b513-0726b7dbce91\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467422 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-oauth-serving-cert\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467438 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e389c321-1fee-4503-8f27-ce2eb2dd76c3-node-pullsecrets\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467456 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/de8e8906-1b66-4e41-8881-513f911a7407-etcd-service-ca\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467473 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1ca9c41a-969e-4475-8487-78f6f4d8e6b8-proxy-tls\") pod \"machine-config-controller-84d6567774-vfk86\" (UID: \"1ca9c41a-969e-4475-8487-78f6f4d8e6b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467489 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e389c321-1fee-4503-8f27-ce2eb2dd76c3-serving-cert\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467508 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467541 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-n9pnl\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467560 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47cfb7d6-3723-4269-95ec-f6a512335387-metrics-certs\") pod \"router-default-5444994796-qtdmw\" (UID: \"47cfb7d6-3723-4269-95ec-f6a512335387\") " pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467576 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f498d61-1987-48a1-82d2-14ea6db95c84-serving-cert\") pod \"controller-manager-879f6c89f-n9pnl\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467593 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/62e1b21d-36c2-4047-8913-8fc45e0401d7-metrics-tls\") pod \"dns-operator-744455d44c-vrmnl\" (UID: \"62e1b21d-36c2-4047-8913-8fc45e0401d7\") " pod="openshift-dns-operator/dns-operator-744455d44c-vrmnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467616 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d9076471-a7f4-437b-82fc-5384b61cc048-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mm2tg\" (UID: \"d9076471-a7f4-437b-82fc-5384b61cc048\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467632 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467651 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pf6z\" (UniqueName: \"kubernetes.io/projected/f68aa373-3441-49f3-abc0-d5647d7bd869-kube-api-access-4pf6z\") pod \"control-plane-machine-set-operator-78cbb6b69f-nscjn\" (UID: \"f68aa373-3441-49f3-abc0-d5647d7bd869\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nscjn" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467669 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72b2f\" (UniqueName: \"kubernetes.io/projected/47cfb7d6-3723-4269-95ec-f6a512335387-kube-api-access-72b2f\") pod \"router-default-5444994796-qtdmw\" (UID: \"47cfb7d6-3723-4269-95ec-f6a512335387\") " pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467688 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f59b0400-f54d-47cc-a58f-7a0399fa857d-audit-dir\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467708 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bghh9\" (UniqueName: \"kubernetes.io/projected/1ca9c41a-969e-4475-8487-78f6f4d8e6b8-kube-api-access-bghh9\") pod \"machine-config-controller-84d6567774-vfk86\" (UID: \"1ca9c41a-969e-4475-8487-78f6f4d8e6b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467725 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-service-ca\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467742 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/47cfb7d6-3723-4269-95ec-f6a512335387-default-certificate\") pod \"router-default-5444994796-qtdmw\" (UID: \"47cfb7d6-3723-4269-95ec-f6a512335387\") " pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467773 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e389c321-1fee-4503-8f27-ce2eb2dd76c3-image-import-ca\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467807 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467898 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e389c321-1fee-4503-8f27-ce2eb2dd76c3-etcd-client\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467920 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6bae8a54-7dad-448f-b513-0726b7dbce91-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rhm46\" (UID: \"6bae8a54-7dad-448f-b513-0726b7dbce91\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467945 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de8e8906-1b66-4e41-8881-513f911a7407-config\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467958 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9a8d7330-30c8-4a9c-ae2f-695b2c17deea-trusted-ca\") pod \"ingress-operator-5b745b69d9-p46hd\" (UID: \"9a8d7330-30c8-4a9c-ae2f-695b2c17deea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467966 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e389c321-1fee-4503-8f27-ce2eb2dd76c3-encryption-config\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.467986 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d97e08b3-79e2-434a-8de9-486cac9a7f19-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7mhrd\" (UID: \"d97e08b3-79e2-434a-8de9-486cac9a7f19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mhrd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468012 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/de8e8906-1b66-4e41-8881-513f911a7407-etcd-ca\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468032 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k66vf\" (UniqueName: \"kubernetes.io/projected/109ecad0-ad50-4121-843e-319c8638705f-kube-api-access-k66vf\") pod \"migrator-59844c95c7-gdt2z\" (UID: \"109ecad0-ad50-4121-843e-319c8638705f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gdt2z" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468050 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09f4f7a0-5033-482d-9333-a2255ae9fb8c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncvkn\" (UID: \"09f4f7a0-5033-482d-9333-a2255ae9fb8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncvkn" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468068 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f59b0400-f54d-47cc-a58f-7a0399fa857d-audit-policies\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468075 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468091 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e621a558-87ca-499b-8d99-a3584988ece6-trusted-ca\") pod \"console-operator-58897d9998-25pdm\" (UID: \"e621a558-87ca-499b-8d99-a3584988ece6\") " pod="openshift-console-operator/console-operator-58897d9998-25pdm" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468111 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2b3764a2-69cd-4f7a-ac6e-5250925117e9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mlzfg\" (UID: \"2b3764a2-69cd-4f7a-ac6e-5250925117e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mlzfg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468133 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/2cf82e3e-7ba7-4a3b-adf5-c635561b63d3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-x4r6k\" (UID: \"2cf82e3e-7ba7-4a3b-adf5-c635561b63d3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-x4r6k" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468161 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2295ac8e-aa01-4349-a4a3-843e45df1475-config\") pod \"kube-apiserver-operator-766d6c64bb-7vh2m\" (UID: \"2295ac8e-aa01-4349-a4a3-843e45df1475\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7vh2m" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468181 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-console-config\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468205 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b3764a2-69cd-4f7a-ac6e-5250925117e9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mlzfg\" (UID: \"2b3764a2-69cd-4f7a-ac6e-5250925117e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mlzfg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468227 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e621a558-87ca-499b-8d99-a3584988ece6-config\") pod \"console-operator-58897d9998-25pdm\" (UID: \"e621a558-87ca-499b-8d99-a3584988ece6\") " pod="openshift-console-operator/console-operator-58897d9998-25pdm" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468245 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjxw9\" (UniqueName: \"kubernetes.io/projected/de8e8906-1b66-4e41-8881-513f911a7407-kube-api-access-fjxw9\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468275 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/53029d0b-3ada-4e72-856c-381ab7ae6292-audit-dir\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468320 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1ca9c41a-969e-4475-8487-78f6f4d8e6b8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vfk86\" (UID: \"1ca9c41a-969e-4475-8487-78f6f4d8e6b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468338 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468360 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99680c81-94dd-4f90-9133-1e572dd9f34c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-f92kp\" (UID: \"99680c81-94dd-4f90-9133-1e572dd9f34c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f92kp" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468378 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f59b0400-f54d-47cc-a58f-7a0399fa857d-encryption-config\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468399 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d9076471-a7f4-437b-82fc-5384b61cc048-proxy-tls\") pod \"machine-config-operator-74547568cd-mm2tg\" (UID: \"d9076471-a7f4-437b-82fc-5384b61cc048\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468415 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e389c321-1fee-4503-8f27-ce2eb2dd76c3-audit-dir\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468430 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j2cc\" (UniqueName: \"kubernetes.io/projected/2cf82e3e-7ba7-4a3b-adf5-c635561b63d3-kube-api-access-6j2cc\") pod \"cluster-samples-operator-665b6dd947-x4r6k\" (UID: \"2cf82e3e-7ba7-4a3b-adf5-c635561b63d3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-x4r6k" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468450 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkj69\" (UniqueName: \"kubernetes.io/projected/d97e08b3-79e2-434a-8de9-486cac9a7f19-kube-api-access-nkj69\") pod \"openshift-apiserver-operator-796bbdcf4f-7mhrd\" (UID: \"d97e08b3-79e2-434a-8de9-486cac9a7f19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mhrd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468454 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-client-ca\") pod \"controller-manager-879f6c89f-n9pnl\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468484 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a2629f52-3b35-4707-8a24-53d7bbb70465-console-oauth-config\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468517 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99680c81-94dd-4f90-9133-1e572dd9f34c-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-f92kp\" (UID: \"99680c81-94dd-4f90-9133-1e572dd9f34c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f92kp" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468518 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d9076471-a7f4-437b-82fc-5384b61cc048-images\") pod \"machine-config-operator-74547568cd-mm2tg\" (UID: \"d9076471-a7f4-437b-82fc-5384b61cc048\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468588 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d97e08b3-79e2-434a-8de9-486cac9a7f19-config\") pod \"openshift-apiserver-operator-796bbdcf4f-7mhrd\" (UID: \"d97e08b3-79e2-434a-8de9-486cac9a7f19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mhrd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468593 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjc88\" (UniqueName: \"kubernetes.io/projected/6bae8a54-7dad-448f-b513-0726b7dbce91-kube-api-access-zjc88\") pod \"cluster-image-registry-operator-dc59b4c8b-rhm46\" (UID: \"6bae8a54-7dad-448f-b513-0726b7dbce91\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468694 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f59b0400-f54d-47cc-a58f-7a0399fa857d-audit-dir\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468696 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc6kx\" (UniqueName: \"kubernetes.io/projected/62e1b21d-36c2-4047-8913-8fc45e0401d7-kube-api-access-tc6kx\") pod \"dns-operator-744455d44c-vrmnl\" (UID: \"62e1b21d-36c2-4047-8913-8fc45e0401d7\") " pod="openshift-dns-operator/dns-operator-744455d44c-vrmnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468752 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f59b0400-f54d-47cc-a58f-7a0399fa857d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468785 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/09f4f7a0-5033-482d-9333-a2255ae9fb8c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncvkn\" (UID: \"09f4f7a0-5033-482d-9333-a2255ae9fb8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncvkn" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468839 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9592c228-7f94-4923-818e-2d307bdf8d28-serving-cert\") pod \"openshift-config-operator-7777fb866f-zrqbh\" (UID: \"9592c228-7f94-4923-818e-2d307bdf8d28\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.468897 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9bl4\" (UniqueName: \"kubernetes.io/projected/53029d0b-3ada-4e72-856c-381ab7ae6292-kube-api-access-k9bl4\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.469229 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e389c321-1fee-4503-8f27-ce2eb2dd76c3-etcd-serving-ca\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.469442 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9592c228-7f94-4923-818e-2d307bdf8d28-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zrqbh\" (UID: \"9592c228-7f94-4923-818e-2d307bdf8d28\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.470377 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f59b0400-f54d-47cc-a58f-7a0399fa857d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.470509 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e389c321-1fee-4503-8f27-ce2eb2dd76c3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.470769 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-service-ca\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.470994 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-oauth-serving-cert\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.471047 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e389c321-1fee-4503-8f27-ce2eb2dd76c3-node-pullsecrets\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.471219 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6bae8a54-7dad-448f-b513-0726b7dbce91-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rhm46\" (UID: \"6bae8a54-7dad-448f-b513-0726b7dbce91\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.471578 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-n9pnl\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.471635 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/53029d0b-3ada-4e72-856c-381ab7ae6292-audit-dir\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.471926 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e621a558-87ca-499b-8d99-a3584988ece6-config\") pod \"console-operator-58897d9998-25pdm\" (UID: \"e621a558-87ca-499b-8d99-a3584988ece6\") " pod="openshift-console-operator/console-operator-58897d9998-25pdm" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.471949 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d9076471-a7f4-437b-82fc-5384b61cc048-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mm2tg\" (UID: \"d9076471-a7f4-437b-82fc-5384b61cc048\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.472271 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-trusted-ca-bundle\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.472945 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.473033 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1ca9c41a-969e-4475-8487-78f6f4d8e6b8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vfk86\" (UID: \"1ca9c41a-969e-4475-8487-78f6f4d8e6b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.473044 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f59b0400-f54d-47cc-a58f-7a0399fa857d-audit-policies\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.473178 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-config\") pod \"controller-manager-879f6c89f-n9pnl\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.473279 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e389c321-1fee-4503-8f27-ce2eb2dd76c3-audit-dir\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.473380 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.473399 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.473995 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.474103 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e389c321-1fee-4503-8f27-ce2eb2dd76c3-etcd-client\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.474149 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e621a558-87ca-499b-8d99-a3584988ece6-serving-cert\") pod \"console-operator-58897d9998-25pdm\" (UID: \"e621a558-87ca-499b-8d99-a3584988ece6\") " pod="openshift-console-operator/console-operator-58897d9998-25pdm" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.474253 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e621a558-87ca-499b-8d99-a3584988ece6-trusted-ca\") pod \"console-operator-58897d9998-25pdm\" (UID: \"e621a558-87ca-499b-8d99-a3584988ece6\") " pod="openshift-console-operator/console-operator-58897d9998-25pdm" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.474417 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.474716 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a2629f52-3b35-4707-8a24-53d7bbb70465-console-serving-cert\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.474758 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-console-config\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.474897 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9a8d7330-30c8-4a9c-ae2f-695b2c17deea-metrics-tls\") pod \"ingress-operator-5b745b69d9-p46hd\" (UID: \"9a8d7330-30c8-4a9c-ae2f-695b2c17deea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.475101 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f59b0400-f54d-47cc-a58f-7a0399fa857d-serving-cert\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.475120 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/6bae8a54-7dad-448f-b513-0726b7dbce91-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rhm46\" (UID: \"6bae8a54-7dad-448f-b513-0726b7dbce91\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.475207 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.475453 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.475631 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.475642 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e389c321-1fee-4503-8f27-ce2eb2dd76c3-encryption-config\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.476105 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e389c321-1fee-4503-8f27-ce2eb2dd76c3-serving-cert\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.476220 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.476365 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d97e08b3-79e2-434a-8de9-486cac9a7f19-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-7mhrd\" (UID: \"d97e08b3-79e2-434a-8de9-486cac9a7f19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mhrd" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.476706 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/2cf82e3e-7ba7-4a3b-adf5-c635561b63d3-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-x4r6k\" (UID: \"2cf82e3e-7ba7-4a3b-adf5-c635561b63d3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-x4r6k" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.476921 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f498d61-1987-48a1-82d2-14ea6db95c84-serving-cert\") pod \"controller-manager-879f6c89f-n9pnl\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.477182 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f59b0400-f54d-47cc-a58f-7a0399fa857d-etcd-client\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.477322 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f59b0400-f54d-47cc-a58f-7a0399fa857d-encryption-config\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.478408 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9592c228-7f94-4923-818e-2d307bdf8d28-serving-cert\") pod \"openshift-config-operator-7777fb866f-zrqbh\" (UID: \"9592c228-7f94-4923-818e-2d307bdf8d28\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.478412 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99680c81-94dd-4f90-9133-1e572dd9f34c-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-f92kp\" (UID: \"99680c81-94dd-4f90-9133-1e572dd9f34c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f92kp" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.482397 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a2629f52-3b35-4707-8a24-53d7bbb70465-console-oauth-config\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.490903 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.493388 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/62e1b21d-36c2-4047-8913-8fc45e0401d7-metrics-tls\") pod \"dns-operator-744455d44c-vrmnl\" (UID: \"62e1b21d-36c2-4047-8913-8fc45e0401d7\") " pod="openshift-dns-operator/dns-operator-744455d44c-vrmnl" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.511266 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.529591 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.551412 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.569610 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.571906 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de8e8906-1b66-4e41-8881-513f911a7407-config\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.590039 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.592909 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/de8e8906-1b66-4e41-8881-513f911a7407-etcd-ca\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.610346 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.630930 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.642300 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de8e8906-1b66-4e41-8881-513f911a7407-serving-cert\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.651066 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.651649 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/de8e8906-1b66-4e41-8881-513f911a7407-etcd-service-ca\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.670756 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.691145 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.701776 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/de8e8906-1b66-4e41-8881-513f911a7407-etcd-client\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.711523 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.714550 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2295ac8e-aa01-4349-a4a3-843e45df1475-config\") pod \"kube-apiserver-operator-766d6c64bb-7vh2m\" (UID: \"2295ac8e-aa01-4349-a4a3-843e45df1475\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7vh2m" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.731426 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.750452 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.762234 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2295ac8e-aa01-4349-a4a3-843e45df1475-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7vh2m\" (UID: \"2295ac8e-aa01-4349-a4a3-843e45df1475\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7vh2m" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.770651 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.790616 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.811052 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.819464 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d9076471-a7f4-437b-82fc-5384b61cc048-images\") pod \"machine-config-operator-74547568cd-mm2tg\" (UID: \"d9076471-a7f4-437b-82fc-5384b61cc048\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.831632 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.850573 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.860982 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d9076471-a7f4-437b-82fc-5384b61cc048-proxy-tls\") pod \"machine-config-operator-74547568cd-mm2tg\" (UID: \"d9076471-a7f4-437b-82fc-5384b61cc048\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.870737 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.875880 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1ca9c41a-969e-4475-8487-78f6f4d8e6b8-proxy-tls\") pod \"machine-config-controller-84d6567774-vfk86\" (UID: \"1ca9c41a-969e-4475-8487-78f6f4d8e6b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.890993 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.910985 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.930489 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.951256 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.971159 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 10 18:57:35 crc kubenswrapper[4828]: I1210 18:57:35.990464 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.010285 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.024112 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/47cfb7d6-3723-4269-95ec-f6a512335387-default-certificate\") pod \"router-default-5444994796-qtdmw\" (UID: \"47cfb7d6-3723-4269-95ec-f6a512335387\") " pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.030744 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.040163 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/47cfb7d6-3723-4269-95ec-f6a512335387-stats-auth\") pod \"router-default-5444994796-qtdmw\" (UID: \"47cfb7d6-3723-4269-95ec-f6a512335387\") " pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.051503 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.065452 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/47cfb7d6-3723-4269-95ec-f6a512335387-metrics-certs\") pod \"router-default-5444994796-qtdmw\" (UID: \"47cfb7d6-3723-4269-95ec-f6a512335387\") " pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.071171 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.078428 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47cfb7d6-3723-4269-95ec-f6a512335387-service-ca-bundle\") pod \"router-default-5444994796-qtdmw\" (UID: \"47cfb7d6-3723-4269-95ec-f6a512335387\") " pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.090399 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.111126 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.130862 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.150783 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.159082 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09f4f7a0-5033-482d-9333-a2255ae9fb8c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncvkn\" (UID: \"09f4f7a0-5033-482d-9333-a2255ae9fb8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncvkn" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.170832 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.174613 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09f4f7a0-5033-482d-9333-a2255ae9fb8c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncvkn\" (UID: \"09f4f7a0-5033-482d-9333-a2255ae9fb8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncvkn" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.190420 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.195888 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f68aa373-3441-49f3-abc0-d5647d7bd869-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-nscjn\" (UID: \"f68aa373-3441-49f3-abc0-d5647d7bd869\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nscjn" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.210884 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.231612 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.250701 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.270479 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.281308 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2b3764a2-69cd-4f7a-ac6e-5250925117e9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mlzfg\" (UID: \"2b3764a2-69cd-4f7a-ac6e-5250925117e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mlzfg" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.291299 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.292714 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b3764a2-69cd-4f7a-ac6e-5250925117e9-config\") pod \"kube-controller-manager-operator-78b949d7b-mlzfg\" (UID: \"2b3764a2-69cd-4f7a-ac6e-5250925117e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mlzfg" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.348945 4828 request.go:700] Waited for 1.005983809s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/configmaps?fieldSelector=metadata.name%3Dmarketplace-trusted-ca&limit=500&resourceVersion=0 Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.360520 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.370698 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.391158 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.411452 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.431625 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.450277 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.470329 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.490613 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.510444 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.529924 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.549665 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.569856 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.610627 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.616422 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfz82\" (UniqueName: \"kubernetes.io/projected/477e7e82-6ea9-4b53-8aa6-95507dbcc11c-kube-api-access-bfz82\") pod \"authentication-operator-69f744f599-2kkp7\" (UID: \"477e7e82-6ea9-4b53-8aa6-95507dbcc11c\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.632467 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.649934 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.670773 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.690929 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.737291 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxkql\" (UniqueName: \"kubernetes.io/projected/18526255-d66e-4a2d-85a0-433ce7c59757-kube-api-access-lxkql\") pod \"machine-approver-56656f9798-flkll\" (UID: \"18526255-d66e-4a2d-85a0-433ce7c59757\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.745344 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nwbk\" (UniqueName: \"kubernetes.io/projected/e9d9df65-e6cc-404d-8cf9-f50c566d5595-kube-api-access-4nwbk\") pod \"route-controller-manager-6576b87f9c-j8vl2\" (UID: \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.758313 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.768905 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph6xt\" (UniqueName: \"kubernetes.io/projected/de5e17c4-b5df-48a1-898c-436e2c0923c3-kube-api-access-ph6xt\") pod \"machine-api-operator-5694c8668f-gspnq\" (UID: \"de5e17c4-b5df-48a1-898c-436e2c0923c3\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.770739 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.770918 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.789573 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.799521 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.811468 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.831219 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.851156 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.870864 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.893216 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.912100 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.931148 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.951543 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.972876 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gspnq"] Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.973317 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 10 18:57:36 crc kubenswrapper[4828]: I1210 18:57:36.990175 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.010308 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.016500 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2kkp7"] Dec 10 18:57:37 crc kubenswrapper[4828]: W1210 18:57:37.026158 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod477e7e82_6ea9_4b53_8aa6_95507dbcc11c.slice/crio-fb6cf18b92fb8671a08b325a04be319b02ee0ab9a51fd6a6a4eaaa376aec9da3 WatchSource:0}: Error finding container fb6cf18b92fb8671a08b325a04be319b02ee0ab9a51fd6a6a4eaaa376aec9da3: Status 404 returned error can't find the container with id fb6cf18b92fb8671a08b325a04be319b02ee0ab9a51fd6a6a4eaaa376aec9da3 Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.031282 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.044668 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.050237 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.070960 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.090401 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.110525 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.130711 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.149788 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.169784 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.190362 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.210658 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.229953 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.249528 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.270099 4828 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.290088 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.310717 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.345381 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9a8d7330-30c8-4a9c-ae2f-695b2c17deea-bound-sa-token\") pod \"ingress-operator-5b745b69d9-p46hd\" (UID: \"9a8d7330-30c8-4a9c-ae2f-695b2c17deea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.348981 4828 request.go:700] Waited for 1.883558645s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/serviceaccounts/machine-config-operator/token Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.374929 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bq986\" (UniqueName: \"kubernetes.io/projected/d9076471-a7f4-437b-82fc-5384b61cc048-kube-api-access-bq986\") pod \"machine-config-operator-74547568cd-mm2tg\" (UID: \"d9076471-a7f4-437b-82fc-5384b61cc048\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.394562 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6bae8a54-7dad-448f-b513-0726b7dbce91-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rhm46\" (UID: \"6bae8a54-7dad-448f-b513-0726b7dbce91\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.418582 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bngnt\" (UniqueName: \"kubernetes.io/projected/510cc49c-3ed7-4998-9d29-c50a7ab86337-kube-api-access-bngnt\") pod \"downloads-7954f5f757-s4zkd\" (UID: \"510cc49c-3ed7-4998-9d29-c50a7ab86337\") " pod="openshift-console/downloads-7954f5f757-s4zkd" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.444760 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzflq\" (UniqueName: \"kubernetes.io/projected/e621a558-87ca-499b-8d99-a3584988ece6-kube-api-access-xzflq\") pod \"console-operator-58897d9998-25pdm\" (UID: \"e621a558-87ca-499b-8d99-a3584988ece6\") " pod="openshift-console-operator/console-operator-58897d9998-25pdm" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.445850 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2295ac8e-aa01-4349-a4a3-843e45df1475-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7vh2m\" (UID: \"2295ac8e-aa01-4349-a4a3-843e45df1475\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7vh2m" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.463969 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" event={"ID":"477e7e82-6ea9-4b53-8aa6-95507dbcc11c","Type":"ContainerStarted","Data":"fb6cf18b92fb8671a08b325a04be319b02ee0ab9a51fd6a6a4eaaa376aec9da3"} Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.466366 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" event={"ID":"de5e17c4-b5df-48a1-898c-436e2c0923c3","Type":"ContainerStarted","Data":"2a1f4780645c615e8a3d832001e291e195be861fed4aa9b19cfd8e161b8cf382"} Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.466399 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" event={"ID":"de5e17c4-b5df-48a1-898c-436e2c0923c3","Type":"ContainerStarted","Data":"b669a796dc88ffb050300482ccc422abb5f6c5b823f61f2c44a7280bfc74dd0a"} Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.468089 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" event={"ID":"18526255-d66e-4a2d-85a0-433ce7c59757","Type":"ContainerStarted","Data":"6a92a6765bca7aaf00b63ad2e85becc0d26e55abf75056931ddd5d1611369d8d"} Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.468118 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" event={"ID":"18526255-d66e-4a2d-85a0-433ce7c59757","Type":"ContainerStarted","Data":"c637aa574968b05279b847f03a6d088d8e95e5c51af406e8196f92dc786c694a"} Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.470213 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pf6z\" (UniqueName: \"kubernetes.io/projected/f68aa373-3441-49f3-abc0-d5647d7bd869-kube-api-access-4pf6z\") pod \"control-plane-machine-set-operator-78cbb6b69f-nscjn\" (UID: \"f68aa373-3441-49f3-abc0-d5647d7bd869\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nscjn" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.489347 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjc88\" (UniqueName: \"kubernetes.io/projected/6bae8a54-7dad-448f-b513-0726b7dbce91-kube-api-access-zjc88\") pod \"cluster-image-registry-operator-dc59b4c8b-rhm46\" (UID: \"6bae8a54-7dad-448f-b513-0726b7dbce91\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.504743 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.512594 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72b2f\" (UniqueName: \"kubernetes.io/projected/47cfb7d6-3723-4269-95ec-f6a512335387-kube-api-access-72b2f\") pod \"router-default-5444994796-qtdmw\" (UID: \"47cfb7d6-3723-4269-95ec-f6a512335387\") " pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.536381 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bghh9\" (UniqueName: \"kubernetes.io/projected/1ca9c41a-969e-4475-8487-78f6f4d8e6b8-kube-api-access-bghh9\") pod \"machine-config-controller-84d6567774-vfk86\" (UID: \"1ca9c41a-969e-4475-8487-78f6f4d8e6b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.538042 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-s4zkd" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.546311 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-25pdm" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.549559 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs4d5\" (UniqueName: \"kubernetes.io/projected/9592c228-7f94-4923-818e-2d307bdf8d28-kube-api-access-rs4d5\") pod \"openshift-config-operator-7777fb866f-zrqbh\" (UID: \"9592c228-7f94-4923-818e-2d307bdf8d28\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.566876 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq5cn\" (UniqueName: \"kubernetes.io/projected/5f498d61-1987-48a1-82d2-14ea6db95c84-kube-api-access-fq5cn\") pod \"controller-manager-879f6c89f-n9pnl\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.584504 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9bl4\" (UniqueName: \"kubernetes.io/projected/53029d0b-3ada-4e72-856c-381ab7ae6292-kube-api-access-k9bl4\") pod \"oauth-openshift-558db77b4-hgsxd\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.611343 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgvzq\" (UniqueName: \"kubernetes.io/projected/a2629f52-3b35-4707-8a24-53d7bbb70465-kube-api-access-vgvzq\") pod \"console-f9d7485db-m4fkj\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.618507 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7vh2m" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.621723 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2"] Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.625590 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc6kx\" (UniqueName: \"kubernetes.io/projected/62e1b21d-36c2-4047-8913-8fc45e0401d7-kube-api-access-tc6kx\") pod \"dns-operator-744455d44c-vrmnl\" (UID: \"62e1b21d-36c2-4047-8913-8fc45e0401d7\") " pod="openshift-dns-operator/dns-operator-744455d44c-vrmnl" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.633251 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.641906 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.644914 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/09f4f7a0-5033-482d-9333-a2255ae9fb8c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ncvkn\" (UID: \"09f4f7a0-5033-482d-9333-a2255ae9fb8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncvkn" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.655001 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.661875 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncvkn" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.668162 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nscjn" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.668215 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llcd7\" (UniqueName: \"kubernetes.io/projected/f59b0400-f54d-47cc-a58f-7a0399fa857d-kube-api-access-llcd7\") pod \"apiserver-7bbb656c7d-8smt2\" (UID: \"f59b0400-f54d-47cc-a58f-7a0399fa857d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.687011 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nltk\" (UniqueName: \"kubernetes.io/projected/99680c81-94dd-4f90-9133-1e572dd9f34c-kube-api-access-6nltk\") pod \"openshift-controller-manager-operator-756b6f6bc6-f92kp\" (UID: \"99680c81-94dd-4f90-9133-1e572dd9f34c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f92kp" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.711180 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snj5z\" (UniqueName: \"kubernetes.io/projected/9a8d7330-30c8-4a9c-ae2f-695b2c17deea-kube-api-access-snj5z\") pod \"ingress-operator-5b745b69d9-p46hd\" (UID: \"9a8d7330-30c8-4a9c-ae2f-695b2c17deea\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.726418 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzz6l\" (UniqueName: \"kubernetes.io/projected/e389c321-1fee-4503-8f27-ce2eb2dd76c3-kube-api-access-zzz6l\") pod \"apiserver-76f77b778f-8xs5l\" (UID: \"e389c321-1fee-4503-8f27-ce2eb2dd76c3\") " pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.729655 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.739087 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f92kp" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.742479 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46"] Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.757622 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.762080 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjxw9\" (UniqueName: \"kubernetes.io/projected/de8e8906-1b66-4e41-8881-513f911a7407-kube-api-access-fjxw9\") pod \"etcd-operator-b45778765-vfsz7\" (UID: \"de8e8906-1b66-4e41-8881-513f911a7407\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.767193 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.775300 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k66vf\" (UniqueName: \"kubernetes.io/projected/109ecad0-ad50-4121-843e-319c8638705f-kube-api-access-k66vf\") pod \"migrator-59844c95c7-gdt2z\" (UID: \"109ecad0-ad50-4121-843e-319c8638705f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gdt2z" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.779918 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-s4zkd"] Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.779965 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-25pdm"] Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.790995 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2b3764a2-69cd-4f7a-ac6e-5250925117e9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mlzfg\" (UID: \"2b3764a2-69cd-4f7a-ac6e-5250925117e9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mlzfg" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.793698 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.808781 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j2cc\" (UniqueName: \"kubernetes.io/projected/2cf82e3e-7ba7-4a3b-adf5-c635561b63d3-kube-api-access-6j2cc\") pod \"cluster-samples-operator-665b6dd947-x4r6k\" (UID: \"2cf82e3e-7ba7-4a3b-adf5-c635561b63d3\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-x4r6k" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.814295 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.825294 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkj69\" (UniqueName: \"kubernetes.io/projected/d97e08b3-79e2-434a-8de9-486cac9a7f19-kube-api-access-nkj69\") pod \"openshift-apiserver-operator-796bbdcf4f-7mhrd\" (UID: \"d97e08b3-79e2-434a-8de9-486cac9a7f19\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mhrd" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.829063 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" Dec 10 18:57:37 crc kubenswrapper[4828]: W1210 18:57:37.840166 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode621a558_87ca_499b_8d99_a3584988ece6.slice/crio-d52cda44f220293e2441e2ee733dd8795b1ebef41c459d369bf5f090fc18be3b WatchSource:0}: Error finding container d52cda44f220293e2441e2ee733dd8795b1ebef41c459d369bf5f090fc18be3b: Status 404 returned error can't find the container with id d52cda44f220293e2441e2ee733dd8795b1ebef41c459d369bf5f090fc18be3b Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.854102 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.869494 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-x4r6k" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.907617 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1e72c425-11b5-4fd4-9f71-9df4d4489475-ca-trust-extracted\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.907678 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-registry-tls\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.907726 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1e72c425-11b5-4fd4-9f71-9df4d4489475-installation-pull-secrets\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.907755 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1e72c425-11b5-4fd4-9f71-9df4d4489475-registry-certificates\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.907790 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.907833 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1e72c425-11b5-4fd4-9f71-9df4d4489475-trusted-ca\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.907870 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-bound-sa-token\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.907898 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmz86\" (UniqueName: \"kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-kube-api-access-cmz86\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:37 crc kubenswrapper[4828]: E1210 18:57:37.908288 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:38.408273447 +0000 UTC m=+138.918884452 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.909919 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.925260 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-vrmnl" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.948258 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gdt2z" Dec 10 18:57:37 crc kubenswrapper[4828]: I1210 18:57:37.977776 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mlzfg" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.008684 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.008941 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghn4r\" (UniqueName: \"kubernetes.io/projected/493b4c81-95b7-44b8-a17b-f103d9602533-kube-api-access-ghn4r\") pod \"service-ca-operator-777779d784-fv8kj\" (UID: \"493b4c81-95b7-44b8-a17b-f103d9602533\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fv8kj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009012 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g96tf\" (UniqueName: \"kubernetes.io/projected/c78748d2-622b-4cab-a462-793db246bc8e-kube-api-access-g96tf\") pod \"multus-admission-controller-857f4d67dd-mqr27\" (UID: \"c78748d2-622b-4cab-a462-793db246bc8e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mqr27" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009058 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-registry-tls\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009085 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c5bd6649-abb0-4898-a73f-efceb62d2cb9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-w7n9w\" (UID: \"c5bd6649-abb0-4898-a73f-efceb62d2cb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009121 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2f1230e1-443a-43c5-bf94-40ef6201a476-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9jzvz\" (UID: \"2f1230e1-443a-43c5-bf94-40ef6201a476\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9jzvz" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009155 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-socket-dir\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009180 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e8c0e685-9945-49df-818a-02daf5bb6099-profile-collector-cert\") pod \"olm-operator-6b444d44fb-jjrlf\" (UID: \"e8c0e685-9945-49df-818a-02daf5bb6099\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009234 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw5rg\" (UniqueName: \"kubernetes.io/projected/01750079-2ca2-4831-be15-a6d273d78680-kube-api-access-fw5rg\") pod \"service-ca-9c57cc56f-wd865\" (UID: \"01750079-2ca2-4831-be15-a6d273d78680\") " pod="openshift-service-ca/service-ca-9c57cc56f-wd865" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009287 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20d80266-84dc-4dee-83b6-3ccd9148fbca-secret-volume\") pod \"collect-profiles-29423205-s2lmj\" (UID: \"20d80266-84dc-4dee-83b6-3ccd9148fbca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009305 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/48ee3baf-5f6c-48ed-b7f5-b82a4679f67f-apiservice-cert\") pod \"packageserver-d55dfcdfc-c8h8c\" (UID: \"48ee3baf-5f6c-48ed-b7f5-b82a4679f67f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009319 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/493b4c81-95b7-44b8-a17b-f103d9602533-serving-cert\") pod \"service-ca-operator-777779d784-fv8kj\" (UID: \"493b4c81-95b7-44b8-a17b-f103d9602533\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fv8kj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009347 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1e72c425-11b5-4fd4-9f71-9df4d4489475-installation-pull-secrets\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009384 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-csi-data-dir\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009401 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpjsv\" (UniqueName: \"kubernetes.io/projected/b0dc0bfb-c602-4e86-8e7f-9156696bdaf0-kube-api-access-wpjsv\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7x5g\" (UID: \"b0dc0bfb-c602-4e86-8e7f-9156696bdaf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7x5g" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009442 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1e72c425-11b5-4fd4-9f71-9df4d4489475-registry-certificates\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009460 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qvwq\" (UniqueName: \"kubernetes.io/projected/c5bd6649-abb0-4898-a73f-efceb62d2cb9-kube-api-access-4qvwq\") pod \"marketplace-operator-79b997595-w7n9w\" (UID: \"c5bd6649-abb0-4898-a73f-efceb62d2cb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009486 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wbfp\" (UniqueName: \"kubernetes.io/projected/e8c0e685-9945-49df-818a-02daf5bb6099-kube-api-access-6wbfp\") pod \"olm-operator-6b444d44fb-jjrlf\" (UID: \"e8c0e685-9945-49df-818a-02daf5bb6099\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009511 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0dc0bfb-c602-4e86-8e7f-9156696bdaf0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7x5g\" (UID: \"b0dc0bfb-c602-4e86-8e7f-9156696bdaf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7x5g" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009527 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/76b89fc6-2d65-415f-9680-f7a30d1217f1-profile-collector-cert\") pod \"catalog-operator-68c6474976-kmgkf\" (UID: \"76b89fc6-2d65-415f-9680-f7a30d1217f1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009599 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/be16c9ea-01fe-4723-94f5-52a1f7a23cd9-metrics-tls\") pod \"dns-default-hgg7d\" (UID: \"be16c9ea-01fe-4723-94f5-52a1f7a23cd9\") " pod="openshift-dns/dns-default-hgg7d" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009624 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1e72c425-11b5-4fd4-9f71-9df4d4489475-trusted-ca\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009649 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/417523fb-3ff4-4747-a387-e9d2021d6006-cert\") pod \"ingress-canary-bwnph\" (UID: \"417523fb-3ff4-4747-a387-e9d2021d6006\") " pod="openshift-ingress-canary/ingress-canary-bwnph" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009664 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/493b4c81-95b7-44b8-a17b-f103d9602533-config\") pod \"service-ca-operator-777779d784-fv8kj\" (UID: \"493b4c81-95b7-44b8-a17b-f103d9602533\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fv8kj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009701 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jsfd\" (UniqueName: \"kubernetes.io/projected/417523fb-3ff4-4747-a387-e9d2021d6006-kube-api-access-7jsfd\") pod \"ingress-canary-bwnph\" (UID: \"417523fb-3ff4-4747-a387-e9d2021d6006\") " pod="openshift-ingress-canary/ingress-canary-bwnph" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009719 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-bound-sa-token\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009733 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/01750079-2ca2-4831-be15-a6d273d78680-signing-cabundle\") pod \"service-ca-9c57cc56f-wd865\" (UID: \"01750079-2ca2-4831-be15-a6d273d78680\") " pod="openshift-service-ca/service-ca-9c57cc56f-wd865" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009838 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-registration-dir\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009856 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9c95e115-3ad7-46dc-b9cd-6a86f155512a-node-bootstrap-token\") pod \"machine-config-server-22lss\" (UID: \"9c95e115-3ad7-46dc-b9cd-6a86f155512a\") " pod="openshift-machine-config-operator/machine-config-server-22lss" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009880 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnhj6\" (UniqueName: \"kubernetes.io/projected/9c95e115-3ad7-46dc-b9cd-6a86f155512a-kube-api-access-cnhj6\") pod \"machine-config-server-22lss\" (UID: \"9c95e115-3ad7-46dc-b9cd-6a86f155512a\") " pod="openshift-machine-config-operator/machine-config-server-22lss" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009923 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkslb\" (UniqueName: \"kubernetes.io/projected/2f1230e1-443a-43c5-bf94-40ef6201a476-kube-api-access-qkslb\") pod \"package-server-manager-789f6589d5-9jzvz\" (UID: \"2f1230e1-443a-43c5-bf94-40ef6201a476\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9jzvz" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009949 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmz86\" (UniqueName: \"kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-kube-api-access-cmz86\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.009965 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/76b89fc6-2d65-415f-9680-f7a30d1217f1-srv-cert\") pod \"catalog-operator-68c6474976-kmgkf\" (UID: \"76b89fc6-2d65-415f-9680-f7a30d1217f1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010005 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97pp7\" (UniqueName: \"kubernetes.io/projected/be16c9ea-01fe-4723-94f5-52a1f7a23cd9-kube-api-access-97pp7\") pod \"dns-default-hgg7d\" (UID: \"be16c9ea-01fe-4723-94f5-52a1f7a23cd9\") " pod="openshift-dns/dns-default-hgg7d" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010086 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20d80266-84dc-4dee-83b6-3ccd9148fbca-config-volume\") pod \"collect-profiles-29423205-s2lmj\" (UID: \"20d80266-84dc-4dee-83b6-3ccd9148fbca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010163 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n74mm\" (UniqueName: \"kubernetes.io/projected/76b89fc6-2d65-415f-9680-f7a30d1217f1-kube-api-access-n74mm\") pod \"catalog-operator-68c6474976-kmgkf\" (UID: \"76b89fc6-2d65-415f-9680-f7a30d1217f1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010190 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9c95e115-3ad7-46dc-b9cd-6a86f155512a-certs\") pod \"machine-config-server-22lss\" (UID: \"9c95e115-3ad7-46dc-b9cd-6a86f155512a\") " pod="openshift-machine-config-operator/machine-config-server-22lss" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010205 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e8c0e685-9945-49df-818a-02daf5bb6099-srv-cert\") pod \"olm-operator-6b444d44fb-jjrlf\" (UID: \"e8c0e685-9945-49df-818a-02daf5bb6099\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010254 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xr8q\" (UniqueName: \"kubernetes.io/projected/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-kube-api-access-7xr8q\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010277 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/48ee3baf-5f6c-48ed-b7f5-b82a4679f67f-webhook-cert\") pod \"packageserver-d55dfcdfc-c8h8c\" (UID: \"48ee3baf-5f6c-48ed-b7f5-b82a4679f67f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010298 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be16c9ea-01fe-4723-94f5-52a1f7a23cd9-config-volume\") pod \"dns-default-hgg7d\" (UID: \"be16c9ea-01fe-4723-94f5-52a1f7a23cd9\") " pod="openshift-dns/dns-default-hgg7d" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010357 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1e72c425-11b5-4fd4-9f71-9df4d4489475-ca-trust-extracted\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010383 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/01750079-2ca2-4831-be15-a6d273d78680-signing-key\") pod \"service-ca-9c57cc56f-wd865\" (UID: \"01750079-2ca2-4831-be15-a6d273d78680\") " pod="openshift-service-ca/service-ca-9c57cc56f-wd865" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010432 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tnmp\" (UniqueName: \"kubernetes.io/projected/20d80266-84dc-4dee-83b6-3ccd9148fbca-kube-api-access-4tnmp\") pod \"collect-profiles-29423205-s2lmj\" (UID: \"20d80266-84dc-4dee-83b6-3ccd9148fbca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010474 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mphr\" (UniqueName: \"kubernetes.io/projected/48ee3baf-5f6c-48ed-b7f5-b82a4679f67f-kube-api-access-8mphr\") pod \"packageserver-d55dfcdfc-c8h8c\" (UID: \"48ee3baf-5f6c-48ed-b7f5-b82a4679f67f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010522 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-mountpoint-dir\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010537 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/48ee3baf-5f6c-48ed-b7f5-b82a4679f67f-tmpfs\") pod \"packageserver-d55dfcdfc-c8h8c\" (UID: \"48ee3baf-5f6c-48ed-b7f5-b82a4679f67f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010581 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5bd6649-abb0-4898-a73f-efceb62d2cb9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-w7n9w\" (UID: \"c5bd6649-abb0-4898-a73f-efceb62d2cb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010643 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0dc0bfb-c602-4e86-8e7f-9156696bdaf0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7x5g\" (UID: \"b0dc0bfb-c602-4e86-8e7f-9156696bdaf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7x5g" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010660 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-plugins-dir\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.010677 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c78748d2-622b-4cab-a462-793db246bc8e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mqr27\" (UID: \"c78748d2-622b-4cab-a462-793db246bc8e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mqr27" Dec 10 18:57:38 crc kubenswrapper[4828]: E1210 18:57:38.011493 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:38.511476859 +0000 UTC m=+139.022087854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.014622 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1e72c425-11b5-4fd4-9f71-9df4d4489475-ca-trust-extracted\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.020192 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1e72c425-11b5-4fd4-9f71-9df4d4489475-trusted-ca\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.020595 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1e72c425-11b5-4fd4-9f71-9df4d4489475-registry-certificates\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.036770 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-registry-tls\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.038182 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1e72c425-11b5-4fd4-9f71-9df4d4489475-installation-pull-secrets\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.060481 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-bound-sa-token\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.088325 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmz86\" (UniqueName: \"kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-kube-api-access-cmz86\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.113902 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-socket-dir\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.113941 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e8c0e685-9945-49df-818a-02daf5bb6099-profile-collector-cert\") pod \"olm-operator-6b444d44fb-jjrlf\" (UID: \"e8c0e685-9945-49df-818a-02daf5bb6099\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.113959 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw5rg\" (UniqueName: \"kubernetes.io/projected/01750079-2ca2-4831-be15-a6d273d78680-kube-api-access-fw5rg\") pod \"service-ca-9c57cc56f-wd865\" (UID: \"01750079-2ca2-4831-be15-a6d273d78680\") " pod="openshift-service-ca/service-ca-9c57cc56f-wd865" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.113978 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20d80266-84dc-4dee-83b6-3ccd9148fbca-secret-volume\") pod \"collect-profiles-29423205-s2lmj\" (UID: \"20d80266-84dc-4dee-83b6-3ccd9148fbca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.113992 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/48ee3baf-5f6c-48ed-b7f5-b82a4679f67f-apiservice-cert\") pod \"packageserver-d55dfcdfc-c8h8c\" (UID: \"48ee3baf-5f6c-48ed-b7f5-b82a4679f67f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114007 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/493b4c81-95b7-44b8-a17b-f103d9602533-serving-cert\") pod \"service-ca-operator-777779d784-fv8kj\" (UID: \"493b4c81-95b7-44b8-a17b-f103d9602533\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fv8kj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114036 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-csi-data-dir\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114050 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpjsv\" (UniqueName: \"kubernetes.io/projected/b0dc0bfb-c602-4e86-8e7f-9156696bdaf0-kube-api-access-wpjsv\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7x5g\" (UID: \"b0dc0bfb-c602-4e86-8e7f-9156696bdaf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7x5g" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114069 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qvwq\" (UniqueName: \"kubernetes.io/projected/c5bd6649-abb0-4898-a73f-efceb62d2cb9-kube-api-access-4qvwq\") pod \"marketplace-operator-79b997595-w7n9w\" (UID: \"c5bd6649-abb0-4898-a73f-efceb62d2cb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114085 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wbfp\" (UniqueName: \"kubernetes.io/projected/e8c0e685-9945-49df-818a-02daf5bb6099-kube-api-access-6wbfp\") pod \"olm-operator-6b444d44fb-jjrlf\" (UID: \"e8c0e685-9945-49df-818a-02daf5bb6099\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114101 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0dc0bfb-c602-4e86-8e7f-9156696bdaf0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7x5g\" (UID: \"b0dc0bfb-c602-4e86-8e7f-9156696bdaf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7x5g" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114115 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/76b89fc6-2d65-415f-9680-f7a30d1217f1-profile-collector-cert\") pod \"catalog-operator-68c6474976-kmgkf\" (UID: \"76b89fc6-2d65-415f-9680-f7a30d1217f1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114137 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114156 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/be16c9ea-01fe-4723-94f5-52a1f7a23cd9-metrics-tls\") pod \"dns-default-hgg7d\" (UID: \"be16c9ea-01fe-4723-94f5-52a1f7a23cd9\") " pod="openshift-dns/dns-default-hgg7d" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114173 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/417523fb-3ff4-4747-a387-e9d2021d6006-cert\") pod \"ingress-canary-bwnph\" (UID: \"417523fb-3ff4-4747-a387-e9d2021d6006\") " pod="openshift-ingress-canary/ingress-canary-bwnph" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114186 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/493b4c81-95b7-44b8-a17b-f103d9602533-config\") pod \"service-ca-operator-777779d784-fv8kj\" (UID: \"493b4c81-95b7-44b8-a17b-f103d9602533\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fv8kj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114207 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jsfd\" (UniqueName: \"kubernetes.io/projected/417523fb-3ff4-4747-a387-e9d2021d6006-kube-api-access-7jsfd\") pod \"ingress-canary-bwnph\" (UID: \"417523fb-3ff4-4747-a387-e9d2021d6006\") " pod="openshift-ingress-canary/ingress-canary-bwnph" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114226 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/01750079-2ca2-4831-be15-a6d273d78680-signing-cabundle\") pod \"service-ca-9c57cc56f-wd865\" (UID: \"01750079-2ca2-4831-be15-a6d273d78680\") " pod="openshift-service-ca/service-ca-9c57cc56f-wd865" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114251 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-registration-dir\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114267 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9c95e115-3ad7-46dc-b9cd-6a86f155512a-node-bootstrap-token\") pod \"machine-config-server-22lss\" (UID: \"9c95e115-3ad7-46dc-b9cd-6a86f155512a\") " pod="openshift-machine-config-operator/machine-config-server-22lss" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114281 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnhj6\" (UniqueName: \"kubernetes.io/projected/9c95e115-3ad7-46dc-b9cd-6a86f155512a-kube-api-access-cnhj6\") pod \"machine-config-server-22lss\" (UID: \"9c95e115-3ad7-46dc-b9cd-6a86f155512a\") " pod="openshift-machine-config-operator/machine-config-server-22lss" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114297 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkslb\" (UniqueName: \"kubernetes.io/projected/2f1230e1-443a-43c5-bf94-40ef6201a476-kube-api-access-qkslb\") pod \"package-server-manager-789f6589d5-9jzvz\" (UID: \"2f1230e1-443a-43c5-bf94-40ef6201a476\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9jzvz" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114312 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/76b89fc6-2d65-415f-9680-f7a30d1217f1-srv-cert\") pod \"catalog-operator-68c6474976-kmgkf\" (UID: \"76b89fc6-2d65-415f-9680-f7a30d1217f1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114337 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97pp7\" (UniqueName: \"kubernetes.io/projected/be16c9ea-01fe-4723-94f5-52a1f7a23cd9-kube-api-access-97pp7\") pod \"dns-default-hgg7d\" (UID: \"be16c9ea-01fe-4723-94f5-52a1f7a23cd9\") " pod="openshift-dns/dns-default-hgg7d" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114366 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20d80266-84dc-4dee-83b6-3ccd9148fbca-config-volume\") pod \"collect-profiles-29423205-s2lmj\" (UID: \"20d80266-84dc-4dee-83b6-3ccd9148fbca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114383 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n74mm\" (UniqueName: \"kubernetes.io/projected/76b89fc6-2d65-415f-9680-f7a30d1217f1-kube-api-access-n74mm\") pod \"catalog-operator-68c6474976-kmgkf\" (UID: \"76b89fc6-2d65-415f-9680-f7a30d1217f1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114398 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9c95e115-3ad7-46dc-b9cd-6a86f155512a-certs\") pod \"machine-config-server-22lss\" (UID: \"9c95e115-3ad7-46dc-b9cd-6a86f155512a\") " pod="openshift-machine-config-operator/machine-config-server-22lss" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114413 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e8c0e685-9945-49df-818a-02daf5bb6099-srv-cert\") pod \"olm-operator-6b444d44fb-jjrlf\" (UID: \"e8c0e685-9945-49df-818a-02daf5bb6099\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114431 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xr8q\" (UniqueName: \"kubernetes.io/projected/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-kube-api-access-7xr8q\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114446 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/48ee3baf-5f6c-48ed-b7f5-b82a4679f67f-webhook-cert\") pod \"packageserver-d55dfcdfc-c8h8c\" (UID: \"48ee3baf-5f6c-48ed-b7f5-b82a4679f67f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114462 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be16c9ea-01fe-4723-94f5-52a1f7a23cd9-config-volume\") pod \"dns-default-hgg7d\" (UID: \"be16c9ea-01fe-4723-94f5-52a1f7a23cd9\") " pod="openshift-dns/dns-default-hgg7d" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114479 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/01750079-2ca2-4831-be15-a6d273d78680-signing-key\") pod \"service-ca-9c57cc56f-wd865\" (UID: \"01750079-2ca2-4831-be15-a6d273d78680\") " pod="openshift-service-ca/service-ca-9c57cc56f-wd865" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114497 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tnmp\" (UniqueName: \"kubernetes.io/projected/20d80266-84dc-4dee-83b6-3ccd9148fbca-kube-api-access-4tnmp\") pod \"collect-profiles-29423205-s2lmj\" (UID: \"20d80266-84dc-4dee-83b6-3ccd9148fbca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114515 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mphr\" (UniqueName: \"kubernetes.io/projected/48ee3baf-5f6c-48ed-b7f5-b82a4679f67f-kube-api-access-8mphr\") pod \"packageserver-d55dfcdfc-c8h8c\" (UID: \"48ee3baf-5f6c-48ed-b7f5-b82a4679f67f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114533 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-mountpoint-dir\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114548 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/48ee3baf-5f6c-48ed-b7f5-b82a4679f67f-tmpfs\") pod \"packageserver-d55dfcdfc-c8h8c\" (UID: \"48ee3baf-5f6c-48ed-b7f5-b82a4679f67f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114564 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5bd6649-abb0-4898-a73f-efceb62d2cb9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-w7n9w\" (UID: \"c5bd6649-abb0-4898-a73f-efceb62d2cb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114581 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0dc0bfb-c602-4e86-8e7f-9156696bdaf0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7x5g\" (UID: \"b0dc0bfb-c602-4e86-8e7f-9156696bdaf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7x5g" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114596 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-plugins-dir\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114611 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c78748d2-622b-4cab-a462-793db246bc8e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mqr27\" (UID: \"c78748d2-622b-4cab-a462-793db246bc8e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mqr27" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114628 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghn4r\" (UniqueName: \"kubernetes.io/projected/493b4c81-95b7-44b8-a17b-f103d9602533-kube-api-access-ghn4r\") pod \"service-ca-operator-777779d784-fv8kj\" (UID: \"493b4c81-95b7-44b8-a17b-f103d9602533\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fv8kj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114645 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g96tf\" (UniqueName: \"kubernetes.io/projected/c78748d2-622b-4cab-a462-793db246bc8e-kube-api-access-g96tf\") pod \"multus-admission-controller-857f4d67dd-mqr27\" (UID: \"c78748d2-622b-4cab-a462-793db246bc8e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mqr27" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114664 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c5bd6649-abb0-4898-a73f-efceb62d2cb9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-w7n9w\" (UID: \"c5bd6649-abb0-4898-a73f-efceb62d2cb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.114680 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2f1230e1-443a-43c5-bf94-40ef6201a476-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9jzvz\" (UID: \"2f1230e1-443a-43c5-bf94-40ef6201a476\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9jzvz" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.118370 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-socket-dir\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.119938 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20d80266-84dc-4dee-83b6-3ccd9148fbca-config-volume\") pod \"collect-profiles-29423205-s2lmj\" (UID: \"20d80266-84dc-4dee-83b6-3ccd9148fbca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.123006 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/493b4c81-95b7-44b8-a17b-f103d9602533-config\") pod \"service-ca-operator-777779d784-fv8kj\" (UID: \"493b4c81-95b7-44b8-a17b-f103d9602533\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fv8kj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.123733 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/01750079-2ca2-4831-be15-a6d273d78680-signing-cabundle\") pod \"service-ca-9c57cc56f-wd865\" (UID: \"01750079-2ca2-4831-be15-a6d273d78680\") " pod="openshift-service-ca/service-ca-9c57cc56f-wd865" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.123794 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-registration-dir\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.125765 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mhrd" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.125893 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/be16c9ea-01fe-4723-94f5-52a1f7a23cd9-config-volume\") pod \"dns-default-hgg7d\" (UID: \"be16c9ea-01fe-4723-94f5-52a1f7a23cd9\") " pod="openshift-dns/dns-default-hgg7d" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.125936 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-csi-data-dir\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.126935 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0dc0bfb-c602-4e86-8e7f-9156696bdaf0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7x5g\" (UID: \"b0dc0bfb-c602-4e86-8e7f-9156696bdaf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7x5g" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.127222 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/48ee3baf-5f6c-48ed-b7f5-b82a4679f67f-webhook-cert\") pod \"packageserver-d55dfcdfc-c8h8c\" (UID: \"48ee3baf-5f6c-48ed-b7f5-b82a4679f67f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.127714 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-plugins-dir\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.127792 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-mountpoint-dir\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: E1210 18:57:38.128931 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:38.628915424 +0000 UTC m=+139.139526429 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.129624 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5bd6649-abb0-4898-a73f-efceb62d2cb9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-w7n9w\" (UID: \"c5bd6649-abb0-4898-a73f-efceb62d2cb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.133617 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9c95e115-3ad7-46dc-b9cd-6a86f155512a-certs\") pod \"machine-config-server-22lss\" (UID: \"9c95e115-3ad7-46dc-b9cd-6a86f155512a\") " pod="openshift-machine-config-operator/machine-config-server-22lss" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.134198 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/76b89fc6-2d65-415f-9680-f7a30d1217f1-srv-cert\") pod \"catalog-operator-68c6474976-kmgkf\" (UID: \"76b89fc6-2d65-415f-9680-f7a30d1217f1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.141047 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/48ee3baf-5f6c-48ed-b7f5-b82a4679f67f-tmpfs\") pod \"packageserver-d55dfcdfc-c8h8c\" (UID: \"48ee3baf-5f6c-48ed-b7f5-b82a4679f67f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.141708 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/e8c0e685-9945-49df-818a-02daf5bb6099-profile-collector-cert\") pod \"olm-operator-6b444d44fb-jjrlf\" (UID: \"e8c0e685-9945-49df-818a-02daf5bb6099\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.148279 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.148589 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/01750079-2ca2-4831-be15-a6d273d78680-signing-key\") pod \"service-ca-9c57cc56f-wd865\" (UID: \"01750079-2ca2-4831-be15-a6d273d78680\") " pod="openshift-service-ca/service-ca-9c57cc56f-wd865" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.149224 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0dc0bfb-c602-4e86-8e7f-9156696bdaf0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7x5g\" (UID: \"b0dc0bfb-c602-4e86-8e7f-9156696bdaf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7x5g" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.149223 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/493b4c81-95b7-44b8-a17b-f103d9602533-serving-cert\") pod \"service-ca-operator-777779d784-fv8kj\" (UID: \"493b4c81-95b7-44b8-a17b-f103d9602533\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fv8kj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.149492 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9c95e115-3ad7-46dc-b9cd-6a86f155512a-node-bootstrap-token\") pod \"machine-config-server-22lss\" (UID: \"9c95e115-3ad7-46dc-b9cd-6a86f155512a\") " pod="openshift-machine-config-operator/machine-config-server-22lss" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.149554 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/417523fb-3ff4-4747-a387-e9d2021d6006-cert\") pod \"ingress-canary-bwnph\" (UID: \"417523fb-3ff4-4747-a387-e9d2021d6006\") " pod="openshift-ingress-canary/ingress-canary-bwnph" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.149633 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/48ee3baf-5f6c-48ed-b7f5-b82a4679f67f-apiservice-cert\") pod \"packageserver-d55dfcdfc-c8h8c\" (UID: \"48ee3baf-5f6c-48ed-b7f5-b82a4679f67f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.149664 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/2f1230e1-443a-43c5-bf94-40ef6201a476-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-9jzvz\" (UID: \"2f1230e1-443a-43c5-bf94-40ef6201a476\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9jzvz" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.149895 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20d80266-84dc-4dee-83b6-3ccd9148fbca-secret-volume\") pod \"collect-profiles-29423205-s2lmj\" (UID: \"20d80266-84dc-4dee-83b6-3ccd9148fbca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.154554 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/c78748d2-622b-4cab-a462-793db246bc8e-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-mqr27\" (UID: \"c78748d2-622b-4cab-a462-793db246bc8e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mqr27" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.155019 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/76b89fc6-2d65-415f-9680-f7a30d1217f1-profile-collector-cert\") pod \"catalog-operator-68c6474976-kmgkf\" (UID: \"76b89fc6-2d65-415f-9680-f7a30d1217f1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.156640 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/be16c9ea-01fe-4723-94f5-52a1f7a23cd9-metrics-tls\") pod \"dns-default-hgg7d\" (UID: \"be16c9ea-01fe-4723-94f5-52a1f7a23cd9\") " pod="openshift-dns/dns-default-hgg7d" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.160970 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/e8c0e685-9945-49df-818a-02daf5bb6099-srv-cert\") pod \"olm-operator-6b444d44fb-jjrlf\" (UID: \"e8c0e685-9945-49df-818a-02daf5bb6099\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.168686 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7vh2m"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.170666 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c5bd6649-abb0-4898-a73f-efceb62d2cb9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-w7n9w\" (UID: \"c5bd6649-abb0-4898-a73f-efceb62d2cb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.171673 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mphr\" (UniqueName: \"kubernetes.io/projected/48ee3baf-5f6c-48ed-b7f5-b82a4679f67f-kube-api-access-8mphr\") pod \"packageserver-d55dfcdfc-c8h8c\" (UID: \"48ee3baf-5f6c-48ed-b7f5-b82a4679f67f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.174609 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97pp7\" (UniqueName: \"kubernetes.io/projected/be16c9ea-01fe-4723-94f5-52a1f7a23cd9-kube-api-access-97pp7\") pod \"dns-default-hgg7d\" (UID: \"be16c9ea-01fe-4723-94f5-52a1f7a23cd9\") " pod="openshift-dns/dns-default-hgg7d" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.188646 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n74mm\" (UniqueName: \"kubernetes.io/projected/76b89fc6-2d65-415f-9680-f7a30d1217f1-kube-api-access-n74mm\") pod \"catalog-operator-68c6474976-kmgkf\" (UID: \"76b89fc6-2d65-415f-9680-f7a30d1217f1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.215984 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:38 crc kubenswrapper[4828]: E1210 18:57:38.216431 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:38.71640316 +0000 UTC m=+139.227014165 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.218218 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: E1210 18:57:38.218780 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:38.718764959 +0000 UTC m=+139.229375964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.234244 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jsfd\" (UniqueName: \"kubernetes.io/projected/417523fb-3ff4-4747-a387-e9d2021d6006-kube-api-access-7jsfd\") pod \"ingress-canary-bwnph\" (UID: \"417523fb-3ff4-4747-a387-e9d2021d6006\") " pod="openshift-ingress-canary/ingress-canary-bwnph" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.236089 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nscjn"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.242419 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.262070 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkslb\" (UniqueName: \"kubernetes.io/projected/2f1230e1-443a-43c5-bf94-40ef6201a476-kube-api-access-qkslb\") pod \"package-server-manager-789f6589d5-9jzvz\" (UID: \"2f1230e1-443a-43c5-bf94-40ef6201a476\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9jzvz" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.263917 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncvkn"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.269436 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpjsv\" (UniqueName: \"kubernetes.io/projected/b0dc0bfb-c602-4e86-8e7f-9156696bdaf0-kube-api-access-wpjsv\") pod \"kube-storage-version-migrator-operator-b67b599dd-r7x5g\" (UID: \"b0dc0bfb-c602-4e86-8e7f-9156696bdaf0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7x5g" Dec 10 18:57:38 crc kubenswrapper[4828]: W1210 18:57:38.278915 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9076471_a7f4_437b_82fc_5384b61cc048.slice/crio-dab5fb43cb183a44f6c5722c3e1a3123799a8fdd7e082cb71481185748f782af WatchSource:0}: Error finding container dab5fb43cb183a44f6c5722c3e1a3123799a8fdd7e082cb71481185748f782af: Status 404 returned error can't find the container with id dab5fb43cb183a44f6c5722c3e1a3123799a8fdd7e082cb71481185748f782af Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.296319 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.297712 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qvwq\" (UniqueName: \"kubernetes.io/projected/c5bd6649-abb0-4898-a73f-efceb62d2cb9-kube-api-access-4qvwq\") pod \"marketplace-operator-79b997595-w7n9w\" (UID: \"c5bd6649-abb0-4898-a73f-efceb62d2cb9\") " pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.302858 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7x5g" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.318860 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wbfp\" (UniqueName: \"kubernetes.io/projected/e8c0e685-9945-49df-818a-02daf5bb6099-kube-api-access-6wbfp\") pod \"olm-operator-6b444d44fb-jjrlf\" (UID: \"e8c0e685-9945-49df-818a-02daf5bb6099\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.319205 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:38 crc kubenswrapper[4828]: E1210 18:57:38.319425 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:38.819409489 +0000 UTC m=+139.330020494 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.319496 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: E1210 18:57:38.319833 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:38.819826536 +0000 UTC m=+139.330437541 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.320268 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.329791 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9jzvz" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.352704 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-hgg7d" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.361655 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghn4r\" (UniqueName: \"kubernetes.io/projected/493b4c81-95b7-44b8-a17b-f103d9602533-kube-api-access-ghn4r\") pod \"service-ca-operator-777779d784-fv8kj\" (UID: \"493b4c81-95b7-44b8-a17b-f103d9602533\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fv8kj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.368789 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-bwnph" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.372578 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g96tf\" (UniqueName: \"kubernetes.io/projected/c78748d2-622b-4cab-a462-793db246bc8e-kube-api-access-g96tf\") pod \"multus-admission-controller-857f4d67dd-mqr27\" (UID: \"c78748d2-622b-4cab-a462-793db246bc8e\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-mqr27" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.381570 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw5rg\" (UniqueName: \"kubernetes.io/projected/01750079-2ca2-4831-be15-a6d273d78680-kube-api-access-fw5rg\") pod \"service-ca-9c57cc56f-wd865\" (UID: \"01750079-2ca2-4831-be15-a6d273d78680\") " pod="openshift-service-ca/service-ca-9c57cc56f-wd865" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.406307 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xr8q\" (UniqueName: \"kubernetes.io/projected/3c954830-5a1b-4bc4-ad5b-4b297f2cbc57-kube-api-access-7xr8q\") pod \"csi-hostpathplugin-z2f4h\" (UID: \"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57\") " pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.406898 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnhj6\" (UniqueName: \"kubernetes.io/projected/9c95e115-3ad7-46dc-b9cd-6a86f155512a-kube-api-access-cnhj6\") pod \"machine-config-server-22lss\" (UID: \"9c95e115-3ad7-46dc-b9cd-6a86f155512a\") " pod="openshift-machine-config-operator/machine-config-server-22lss" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.420282 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:38 crc kubenswrapper[4828]: E1210 18:57:38.420662 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:38.92064569 +0000 UTC m=+139.431256695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.427897 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tnmp\" (UniqueName: \"kubernetes.io/projected/20d80266-84dc-4dee-83b6-3ccd9148fbca-kube-api-access-4tnmp\") pod \"collect-profiles-29423205-s2lmj\" (UID: \"20d80266-84dc-4dee-83b6-3ccd9148fbca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.440689 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f92kp"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.469076 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.524625 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: E1210 18:57:38.525030 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:39.025016931 +0000 UTC m=+139.535627936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.527109 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-s4zkd" event={"ID":"510cc49c-3ed7-4998-9d29-c50a7ab86337","Type":"ContainerStarted","Data":"b1a14031fa9acd09ebd3b1783fb20619d123e44a5db76bf7d7d795614e0aedf8"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.527149 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-s4zkd" event={"ID":"510cc49c-3ed7-4998-9d29-c50a7ab86337","Type":"ContainerStarted","Data":"0b34d4e85f1771efeb2126ca8fc451dbe405b4736ff8367d045a50e93ab8c743"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.527889 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-s4zkd" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.532344 4828 patch_prober.go:28] interesting pod/downloads-7954f5f757-s4zkd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.532404 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-s4zkd" podUID="510cc49c-3ed7-4998-9d29-c50a7ab86337" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.535362 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" event={"ID":"e9d9df65-e6cc-404d-8cf9-f50c566d5595","Type":"ContainerStarted","Data":"23203906e6c8e9ab5602b14502fb40b91bdc27164c2394aa378d570566a2f1f8"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.535404 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" event={"ID":"e9d9df65-e6cc-404d-8cf9-f50c566d5595","Type":"ContainerStarted","Data":"abd3339d25d2b6f80c44f5e8a8eb218c62113ad88f460656a676e9e0fd7c902b"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.535900 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.546051 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.549903 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" event={"ID":"de5e17c4-b5df-48a1-898c-436e2c0923c3","Type":"ContainerStarted","Data":"f7ac522a8770aaae1b43a029a8348be8f956c23dcc8d4b55446ccf059ec557b8"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.553968 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46" event={"ID":"6bae8a54-7dad-448f-b513-0726b7dbce91","Type":"ContainerStarted","Data":"8508a6a68cf82bae9e8b57eb0f53b8f550638fa0d0573d04c0203710e37fe809"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.554004 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46" event={"ID":"6bae8a54-7dad-448f-b513-0726b7dbce91","Type":"ContainerStarted","Data":"8dfd5c4515fa7ae3cdd1475b5cac408f2e664ed096b6312d788b39875c08b446"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.569057 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-25pdm" event={"ID":"e621a558-87ca-499b-8d99-a3584988ece6","Type":"ContainerStarted","Data":"2f05538c96f6c6f470dea1cd5ee48dfd4994e191650b303f5f2690ff85b4c5b7"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.569109 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-25pdm" event={"ID":"e621a558-87ca-499b-8d99-a3584988ece6","Type":"ContainerStarted","Data":"d52cda44f220293e2441e2ee733dd8795b1ebef41c459d369bf5f090fc18be3b"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.569733 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-25pdm" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.570732 4828 patch_prober.go:28] interesting pod/console-operator-58897d9998-25pdm container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.570771 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-25pdm" podUID="e621a558-87ca-499b-8d99-a3584988ece6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.574938 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" event={"ID":"d9076471-a7f4-437b-82fc-5384b61cc048","Type":"ContainerStarted","Data":"dab5fb43cb183a44f6c5722c3e1a3123799a8fdd7e082cb71481185748f782af"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.584329 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.591406 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-mqr27" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.602638 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7vh2m" event={"ID":"2295ac8e-aa01-4349-a4a3-843e45df1475","Type":"ContainerStarted","Data":"fc9cd6a120530ea23ff245edea17215d04a2ac00f1d3d57936d20ea45d4fd492"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.611047 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncvkn" event={"ID":"09f4f7a0-5033-482d-9333-a2255ae9fb8c","Type":"ContainerStarted","Data":"05b86da7a965aa32e25c612a74399af2c9fa3559551d80a7e612675d963a77bf"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.611316 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.631016 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:38 crc kubenswrapper[4828]: E1210 18:57:38.632011 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:39.131997865 +0000 UTC m=+139.642608870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.632885 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wd865" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.638451 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.645350 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fv8kj" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.655829 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n9pnl"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.659712 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-vrmnl"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.659923 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-22lss" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.684107 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.691689 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.692000 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-m4fkj"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.694885 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" event={"ID":"477e7e82-6ea9-4b53-8aa6-95507dbcc11c","Type":"ContainerStarted","Data":"f176514874c8b1eb947632df922d2a8e5b792ee4a8284c2e6d9ddc513976e514"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.706114 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8xs5l"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.734550 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: E1210 18:57:38.739571 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:39.239556558 +0000 UTC m=+139.750167563 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.748521 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" event={"ID":"18526255-d66e-4a2d-85a0-433ce7c59757","Type":"ContainerStarted","Data":"38a5f3506d10e589556f2683857c42c1225f4fae9bb68a475906e05fb2748c64"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.775309 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86" event={"ID":"1ca9c41a-969e-4475-8487-78f6f4d8e6b8","Type":"ContainerStarted","Data":"fcc6a04e13ccba562f1da2a959c40293977e3b11472387f8bc25e8d1f7b334af"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.775674 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86" event={"ID":"1ca9c41a-969e-4475-8487-78f6f4d8e6b8","Type":"ContainerStarted","Data":"8c51ed91570cc1b939b6892f7045352b6c32b993f7f964271a1441b2b5bf4c76"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.780428 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nscjn" event={"ID":"f68aa373-3441-49f3-abc0-d5647d7bd869","Type":"ContainerStarted","Data":"dcea3e639186798acb75d36b50112e981bc8dde9c003c5c5b807f9bfd867655a"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.796340 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-qtdmw" event={"ID":"47cfb7d6-3723-4269-95ec-f6a512335387","Type":"ContainerStarted","Data":"f8836bb617fb6c56f7a86207c315a4ade9ea6d61925e44dea0de4413217f99ed"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.796387 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-qtdmw" event={"ID":"47cfb7d6-3723-4269-95ec-f6a512335387","Type":"ContainerStarted","Data":"40b1344715b60b0252a81b0e95dce4cd2354fa92951a6859b46c10a28eb888fb"} Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.802613 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hgsxd"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.819602 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vfsz7"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.819644 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-x4r6k"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.819653 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.828466 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gdt2z"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.842205 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:38 crc kubenswrapper[4828]: E1210 18:57:38.842616 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:39.342577478 +0000 UTC m=+139.853188483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.845103 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.844846 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mlzfg"] Dec 10 18:57:38 crc kubenswrapper[4828]: E1210 18:57:38.845654 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:39.345640868 +0000 UTC m=+139.856251873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.856484 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mhrd"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.889560 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf"] Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.931633 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c"] Dec 10 18:57:38 crc kubenswrapper[4828]: W1210 18:57:38.970489 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53029d0b_3ada_4e72_856c_381ab7ae6292.slice/crio-f2405e5a507e90e1d2a620454b233a57e51405eeba93c2b31ac3cc8288744f8d WatchSource:0}: Error finding container f2405e5a507e90e1d2a620454b233a57e51405eeba93c2b31ac3cc8288744f8d: Status 404 returned error can't find the container with id f2405e5a507e90e1d2a620454b233a57e51405eeba93c2b31ac3cc8288744f8d Dec 10 18:57:38 crc kubenswrapper[4828]: W1210 18:57:38.972147 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod109ecad0_ad50_4121_843e_319c8638705f.slice/crio-601c292a2eff0777ecbc50c47e7a505e692c3f7e3c642248f728525d8d7275d2 WatchSource:0}: Error finding container 601c292a2eff0777ecbc50c47e7a505e692c3f7e3c642248f728525d8d7275d2: Status 404 returned error can't find the container with id 601c292a2eff0777ecbc50c47e7a505e692c3f7e3c642248f728525d8d7275d2 Dec 10 18:57:38 crc kubenswrapper[4828]: I1210 18:57:38.984387 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:38 crc kubenswrapper[4828]: E1210 18:57:38.985056 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:39.48502604 +0000 UTC m=+139.995637045 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.085381 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:39 crc kubenswrapper[4828]: E1210 18:57:39.085757 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:39.585745482 +0000 UTC m=+140.096356477 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:39 crc kubenswrapper[4828]: W1210 18:57:39.167310 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48ee3baf_5f6c_48ed_b7f5_b82a4679f67f.slice/crio-432296c586bad4d08c47828150fc4420a3e93cb3a5a0e8cc5ad7fbb160f25147 WatchSource:0}: Error finding container 432296c586bad4d08c47828150fc4420a3e93cb3a5a0e8cc5ad7fbb160f25147: Status 404 returned error can't find the container with id 432296c586bad4d08c47828150fc4420a3e93cb3a5a0e8cc5ad7fbb160f25147 Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.186185 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:39 crc kubenswrapper[4828]: E1210 18:57:39.186670 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:39.686646777 +0000 UTC m=+140.197257782 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.237080 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-bwnph"] Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.250262 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7x5g"] Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.263653 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-hgg7d"] Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.287461 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:39 crc kubenswrapper[4828]: E1210 18:57:39.287827 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:39.787813636 +0000 UTC m=+140.298424641 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.300888 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9jzvz"] Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.331212 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-gspnq" podStartSLOduration=120.331193513 podStartE2EDuration="2m0.331193513s" podCreationTimestamp="2025-12-10 18:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:39.312617921 +0000 UTC m=+139.823228946" watchObservedRunningTime="2025-12-10 18:57:39.331193513 +0000 UTC m=+139.841804518" Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.331299 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-mqr27"] Dec 10 18:57:39 crc kubenswrapper[4828]: W1210 18:57:39.356331 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod417523fb_3ff4_4747_a387_e9d2021d6006.slice/crio-1e63ce09600e681392a79b58342fed83fe2e789836da45546eef03b96b656cf0 WatchSource:0}: Error finding container 1e63ce09600e681392a79b58342fed83fe2e789836da45546eef03b96b656cf0: Status 404 returned error can't find the container with id 1e63ce09600e681392a79b58342fed83fe2e789836da45546eef03b96b656cf0 Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.387971 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-flkll" podStartSLOduration=121.387934798 podStartE2EDuration="2m1.387934798s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:39.371227586 +0000 UTC m=+139.881838591" watchObservedRunningTime="2025-12-10 18:57:39.387934798 +0000 UTC m=+139.898545803" Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.389500 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:39 crc kubenswrapper[4828]: E1210 18:57:39.389886 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:39.8898693 +0000 UTC m=+140.400480305 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.391962 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w7n9w"] Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.411025 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fv8kj"] Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.495257 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:39 crc kubenswrapper[4828]: E1210 18:57:39.495588 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:39.995577223 +0000 UTC m=+140.506188228 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.501620 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rhm46" podStartSLOduration=121.501607061 podStartE2EDuration="2m1.501607061s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:39.498305178 +0000 UTC m=+140.008916203" watchObservedRunningTime="2025-12-10 18:57:39.501607061 +0000 UTC m=+140.012218066" Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.524873 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf"] Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.547323 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-s4zkd" podStartSLOduration=121.547302815 podStartE2EDuration="2m1.547302815s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:39.547061391 +0000 UTC m=+140.057672406" watchObservedRunningTime="2025-12-10 18:57:39.547302815 +0000 UTC m=+140.057913820" Dec 10 18:57:39 crc kubenswrapper[4828]: W1210 18:57:39.581524 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc78748d2_622b_4cab_a462_793db246bc8e.slice/crio-f08333d945a60c30979a52ba3c272b987ad3b3b2a2985b299cdae22ce3f5a467 WatchSource:0}: Error finding container f08333d945a60c30979a52ba3c272b987ad3b3b2a2985b299cdae22ce3f5a467: Status 404 returned error can't find the container with id f08333d945a60c30979a52ba3c272b987ad3b3b2a2985b299cdae22ce3f5a467 Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.596335 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:39 crc kubenswrapper[4828]: E1210 18:57:39.596838 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:40.096820073 +0000 UTC m=+140.607431078 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:39 crc kubenswrapper[4828]: W1210 18:57:39.601396 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8c0e685_9945_49df_818a_02daf5bb6099.slice/crio-eacd6e8331d4d2512aa4c2498c0b260fd7a8e272c33dc222200111084b94f6da WatchSource:0}: Error finding container eacd6e8331d4d2512aa4c2498c0b260fd7a8e272c33dc222200111084b94f6da: Status 404 returned error can't find the container with id eacd6e8331d4d2512aa4c2498c0b260fd7a8e272c33dc222200111084b94f6da Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.657553 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.665010 4828 patch_prober.go:28] interesting pod/router-default-5444994796-qtdmw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:39 crc kubenswrapper[4828]: [-]has-synced failed: reason withheld Dec 10 18:57:39 crc kubenswrapper[4828]: [+]process-running ok Dec 10 18:57:39 crc kubenswrapper[4828]: healthz check failed Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.665080 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qtdmw" podUID="47cfb7d6-3723-4269-95ec-f6a512335387" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.698730 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:39 crc kubenswrapper[4828]: E1210 18:57:39.699104 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:40.19909175 +0000 UTC m=+140.709702755 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.744544 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-25pdm" podStartSLOduration=121.744517871 podStartE2EDuration="2m1.744517871s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:39.741471811 +0000 UTC m=+140.252082816" watchObservedRunningTime="2025-12-10 18:57:39.744517871 +0000 UTC m=+140.255128876" Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.747712 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-z2f4h"] Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.799727 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:39 crc kubenswrapper[4828]: E1210 18:57:39.799873 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:40.299843522 +0000 UTC m=+140.810454527 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.799988 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:39 crc kubenswrapper[4828]: E1210 18:57:39.800309 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:40.30029921 +0000 UTC m=+140.810910215 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.806365 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86" event={"ID":"1ca9c41a-969e-4475-8487-78f6f4d8e6b8","Type":"ContainerStarted","Data":"6a9e8a77175ad0608b1b3a04184ece971ebea7f4e7168cca2b6eea439bfab2ed"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.808841 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" event={"ID":"53029d0b-3ada-4e72-856c-381ab7ae6292","Type":"ContainerStarted","Data":"f2405e5a507e90e1d2a620454b233a57e51405eeba93c2b31ac3cc8288744f8d"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.810094 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vrmnl" event={"ID":"62e1b21d-36c2-4047-8913-8fc45e0401d7","Type":"ContainerStarted","Data":"a1fcc6546093b974ec8ebf6f9d737fcae4fc22e914fac9cbc37f3199da14a8d9"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.810939 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mqr27" event={"ID":"c78748d2-622b-4cab-a462-793db246bc8e","Type":"ContainerStarted","Data":"f08333d945a60c30979a52ba3c272b987ad3b3b2a2985b299cdae22ce3f5a467"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.812096 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m4fkj" event={"ID":"a2629f52-3b35-4707-8a24-53d7bbb70465","Type":"ContainerStarted","Data":"bd51afe045f8dfaf1d9d491409d261c81fe127e419b1041386caf3c0511323f1"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.813195 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" event={"ID":"f59b0400-f54d-47cc-a58f-7a0399fa857d","Type":"ContainerStarted","Data":"ddbf6426d65b38884f499576d42d86ed71960b520b650a3d9a96f1d19a6a3511"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.814242 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" event={"ID":"c5bd6649-abb0-4898-a73f-efceb62d2cb9","Type":"ContainerStarted","Data":"99f36d3821e1a5847f353ca35068711becb42bfec67bef0fcffed7947792b39e"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.816017 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" event={"ID":"d9076471-a7f4-437b-82fc-5384b61cc048","Type":"ContainerStarted","Data":"c0ab5577b8e3c709fe87491b09e6c8f758eec6dfef0327c7f2838857c041f8cf"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.818949 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f92kp" event={"ID":"99680c81-94dd-4f90-9133-1e572dd9f34c","Type":"ContainerStarted","Data":"91ed7a58a96cc7ddda4f5289820000227f9e7f9f0eb59e5a1ea2139287c7b33c"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.818991 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f92kp" event={"ID":"99680c81-94dd-4f90-9133-1e572dd9f34c","Type":"ContainerStarted","Data":"0ac3781b9eae01d6689cd9ac1b46e96202f1edbb4c2bf815260c5da38d4a3579"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.819858 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf" event={"ID":"76b89fc6-2d65-415f-9680-f7a30d1217f1","Type":"ContainerStarted","Data":"1c69711a1d58667ba65c6f615cd793beac5bfe5d2e8a5f0c337b86de0bad6964"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.820719 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" podStartSLOduration=120.820702712 podStartE2EDuration="2m0.820702712s" podCreationTimestamp="2025-12-10 18:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:39.781695887 +0000 UTC m=+140.292306892" watchObservedRunningTime="2025-12-10 18:57:39.820702712 +0000 UTC m=+140.331313717" Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.821836 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" event={"ID":"e389c321-1fee-4503-8f27-ce2eb2dd76c3","Type":"ContainerStarted","Data":"ab1b99d265bfb5653e645d2b7f1f4a9f06e173f5e8bac0d6aea1d90a2d594428"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.822792 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" event={"ID":"5f498d61-1987-48a1-82d2-14ea6db95c84","Type":"ContainerStarted","Data":"f17edbcc305241a3de574eed4317ad4009f9e1ee99d48ee17a481b55a5ad9822"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.824066 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf" event={"ID":"e8c0e685-9945-49df-818a-02daf5bb6099","Type":"ContainerStarted","Data":"eacd6e8331d4d2512aa4c2498c0b260fd7a8e272c33dc222200111084b94f6da"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.825293 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" event={"ID":"48ee3baf-5f6c-48ed-b7f5-b82a4679f67f","Type":"ContainerStarted","Data":"432296c586bad4d08c47828150fc4420a3e93cb3a5a0e8cc5ad7fbb160f25147"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.826165 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9jzvz" event={"ID":"2f1230e1-443a-43c5-bf94-40ef6201a476","Type":"ContainerStarted","Data":"aa36f3b4fba9111a03eea33ae4bd2c2c176aff8fd4d12545d4bdfc335944adcb"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.827058 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7x5g" event={"ID":"b0dc0bfb-c602-4e86-8e7f-9156696bdaf0","Type":"ContainerStarted","Data":"36b1c6ca6e3df12841c5b9282df3d871a5b129ce7209284420e7eb7d1e074050"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.828148 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" event={"ID":"de8e8906-1b66-4e41-8881-513f911a7407","Type":"ContainerStarted","Data":"c0f61ed6b85e492536d605abcfb7160cf578b51959c863a1d4fcc9d82fa99b23"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.830142 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gdt2z" event={"ID":"109ecad0-ad50-4121-843e-319c8638705f","Type":"ContainerStarted","Data":"601c292a2eff0777ecbc50c47e7a505e692c3f7e3c642248f728525d8d7275d2"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.833419 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mhrd" event={"ID":"d97e08b3-79e2-434a-8de9-486cac9a7f19","Type":"ContainerStarted","Data":"b4f0b3b352d241e1026b48243575b5e98b9f211d670e3592c4d1064868f999e4"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.847182 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" event={"ID":"9a8d7330-30c8-4a9c-ae2f-695b2c17deea","Type":"ContainerStarted","Data":"58a5efa24cf4cbddefd2080adb5855f01439ee1a7be2a38092cfa06e9dcb6e56"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.849625 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mlzfg" event={"ID":"2b3764a2-69cd-4f7a-ac6e-5250925117e9","Type":"ContainerStarted","Data":"539b26ba826925cf3877a5a4f583ba290274dd112b9c7f557feffb86c5f372cb"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.858980 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh" event={"ID":"9592c228-7f94-4923-818e-2d307bdf8d28","Type":"ContainerStarted","Data":"27d7f37784ab8e05bac05c058ec11680c3fb66d05e51502f3bb4bcdf248e5b9a"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.862649 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-2kkp7" podStartSLOduration=121.862636506 podStartE2EDuration="2m1.862636506s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:39.860717225 +0000 UTC m=+140.371328230" watchObservedRunningTime="2025-12-10 18:57:39.862636506 +0000 UTC m=+140.373247511" Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.865082 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fv8kj" event={"ID":"493b4c81-95b7-44b8-a17b-f103d9602533","Type":"ContainerStarted","Data":"8a68e5f07e3210a71e464aff8fffb0bf4f28f9719701fcf13bef5c648129bc22"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.869217 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hgg7d" event={"ID":"be16c9ea-01fe-4723-94f5-52a1f7a23cd9","Type":"ContainerStarted","Data":"7c0376c40613015faa38ea6794c08f6adc6017190532385c90c4744934d34d18"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.870733 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nscjn" event={"ID":"f68aa373-3441-49f3-abc0-d5647d7bd869","Type":"ContainerStarted","Data":"7c02846ea426e44f756b823af74c08cdcd692dd1c58790b9f750e5dcb4c718a1"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.873489 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-bwnph" event={"ID":"417523fb-3ff4-4747-a387-e9d2021d6006","Type":"ContainerStarted","Data":"1e63ce09600e681392a79b58342fed83fe2e789836da45546eef03b96b656cf0"} Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.874478 4828 patch_prober.go:28] interesting pod/downloads-7954f5f757-s4zkd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.874604 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-s4zkd" podUID="510cc49c-3ed7-4998-9d29-c50a7ab86337" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 10 18:57:39 crc kubenswrapper[4828]: I1210 18:57:39.900773 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:39 crc kubenswrapper[4828]: E1210 18:57:39.901163 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:40.401122393 +0000 UTC m=+140.911733548 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.007048 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:40 crc kubenswrapper[4828]: E1210 18:57:40.009339 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:40.509326777 +0000 UTC m=+141.019937782 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.059527 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-25pdm" Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.068750 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-qtdmw" podStartSLOduration=122.068728876 podStartE2EDuration="2m2.068728876s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:40.06716478 +0000 UTC m=+140.577775785" watchObservedRunningTime="2025-12-10 18:57:40.068728876 +0000 UTC m=+140.579339881" Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.108197 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:40 crc kubenswrapper[4828]: E1210 18:57:40.108633 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:40.608617136 +0000 UTC m=+141.119228141 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.176673 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vfk86" podStartSLOduration=122.176646035 podStartE2EDuration="2m2.176646035s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:40.160055355 +0000 UTC m=+140.670666360" watchObservedRunningTime="2025-12-10 18:57:40.176646035 +0000 UTC m=+140.687257040" Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.211526 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:40 crc kubenswrapper[4828]: E1210 18:57:40.211784 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:40.711773688 +0000 UTC m=+141.222384693 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.213747 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj"] Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.239199 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-nscjn" podStartSLOduration=121.239184265 podStartE2EDuration="2m1.239184265s" podCreationTimestamp="2025-12-10 18:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:40.227681197 +0000 UTC m=+140.738292202" watchObservedRunningTime="2025-12-10 18:57:40.239184265 +0000 UTC m=+140.749795270" Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.292434 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wd865"] Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.313079 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:40 crc kubenswrapper[4828]: E1210 18:57:40.313735 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:40.81371859 +0000 UTC m=+141.324329585 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:40 crc kubenswrapper[4828]: W1210 18:57:40.328209 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01750079_2ca2_4831_be15_a6d273d78680.slice/crio-2e7db66a490178ce2f1caf978b963e5eb8d404babc1abb937e807e318cdbf07b WatchSource:0}: Error finding container 2e7db66a490178ce2f1caf978b963e5eb8d404babc1abb937e807e318cdbf07b: Status 404 returned error can't find the container with id 2e7db66a490178ce2f1caf978b963e5eb8d404babc1abb937e807e318cdbf07b Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.414666 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:40 crc kubenswrapper[4828]: E1210 18:57:40.415259 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:40.915248085 +0000 UTC m=+141.425859090 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.515776 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:40 crc kubenswrapper[4828]: E1210 18:57:40.517223 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:41.017203677 +0000 UTC m=+141.527814682 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.617626 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:40 crc kubenswrapper[4828]: E1210 18:57:40.617970 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:41.117958469 +0000 UTC m=+141.628569464 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.657589 4828 patch_prober.go:28] interesting pod/router-default-5444994796-qtdmw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:40 crc kubenswrapper[4828]: [-]has-synced failed: reason withheld Dec 10 18:57:40 crc kubenswrapper[4828]: [+]process-running ok Dec 10 18:57:40 crc kubenswrapper[4828]: healthz check failed Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.657634 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qtdmw" podUID="47cfb7d6-3723-4269-95ec-f6a512335387" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.718056 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:40 crc kubenswrapper[4828]: E1210 18:57:40.718169 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:41.218152613 +0000 UTC m=+141.728763618 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.718224 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:40 crc kubenswrapper[4828]: E1210 18:57:40.718555 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:41.218547829 +0000 UTC m=+141.729158834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.823209 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:40 crc kubenswrapper[4828]: E1210 18:57:40.823710 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:41.323682973 +0000 UTC m=+141.834293978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.823782 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:40 crc kubenswrapper[4828]: E1210 18:57:40.824174 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:41.324160001 +0000 UTC m=+141.834771006 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.928202 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:40 crc kubenswrapper[4828]: E1210 18:57:40.928454 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:41.428438831 +0000 UTC m=+141.939049826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.941088 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mlzfg" event={"ID":"2b3764a2-69cd-4f7a-ac6e-5250925117e9","Type":"ContainerStarted","Data":"e1c1860ca10a9690f841408d55190d458c418b3c16574c887dcd40c21d9251aa"} Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.952788 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" event={"ID":"53029d0b-3ada-4e72-856c-381ab7ae6292","Type":"ContainerStarted","Data":"2551383528f4b0b56017f6748a56555e2a0982d52e02d1b15f00d94866a614be"} Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.964444 4828 generic.go:334] "Generic (PLEG): container finished" podID="e389c321-1fee-4503-8f27-ce2eb2dd76c3" containerID="91b8bc2fb68abd45c428e466d6a0ca4106b3586abc4ca2a737ea47248c1163a7" exitCode=0 Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.964520 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" event={"ID":"e389c321-1fee-4503-8f27-ce2eb2dd76c3","Type":"ContainerDied","Data":"91b8bc2fb68abd45c428e466d6a0ca4106b3586abc4ca2a737ea47248c1163a7"} Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.979301 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gdt2z" event={"ID":"109ecad0-ad50-4121-843e-319c8638705f","Type":"ContainerStarted","Data":"229b8d7f4fd145ef94e2ea19462eda47d7fee607380e19facd7b0b16fa974f69"} Dec 10 18:57:40 crc kubenswrapper[4828]: I1210 18:57:40.991484 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7vh2m" event={"ID":"2295ac8e-aa01-4349-a4a3-843e45df1475","Type":"ContainerStarted","Data":"2a8e44f2e78829f1f1e4c867e6f3c3ea0665f37a6211665e1af2095a18b0d552"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.001580 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" event={"ID":"5f498d61-1987-48a1-82d2-14ea6db95c84","Type":"ContainerStarted","Data":"3c38d62467fe683e8ac5876857183ae6676c9a18fa4b9c7c745e83472657f49a"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.002177 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.011005 4828 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-n9pnl container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.011054 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" podUID="5f498d61-1987-48a1-82d2-14ea6db95c84" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.012882 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" event={"ID":"20d80266-84dc-4dee-83b6-3ccd9148fbca","Type":"ContainerStarted","Data":"9a7006f0f651b1b97ad91a37d2457735d5b58ace10fcdccb370f09464d891442"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.019252 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mhrd" event={"ID":"d97e08b3-79e2-434a-8de9-486cac9a7f19","Type":"ContainerStarted","Data":"d16e74d3d1f10670978c9816daa11acba58e62e22029736f78ff87ae2db24762"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.022937 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m4fkj" event={"ID":"a2629f52-3b35-4707-8a24-53d7bbb70465","Type":"ContainerStarted","Data":"df082dd35197b2e6981657555c2e16a44917cd58def58e905691e03394e43d5c"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.028836 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mlzfg" podStartSLOduration=123.028783507 podStartE2EDuration="2m3.028783507s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:40.973129959 +0000 UTC m=+141.483740964" watchObservedRunningTime="2025-12-10 18:57:41.028783507 +0000 UTC m=+141.539394512" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.029274 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:41 crc kubenswrapper[4828]: E1210 18:57:41.029517 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:41.529504548 +0000 UTC m=+142.040115553 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.055088 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-m4fkj" podStartSLOduration=123.055072015 podStartE2EDuration="2m3.055072015s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:41.053152164 +0000 UTC m=+141.563763179" watchObservedRunningTime="2025-12-10 18:57:41.055072015 +0000 UTC m=+141.565683010" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.073324 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" event={"ID":"d9076471-a7f4-437b-82fc-5384b61cc048","Type":"ContainerStarted","Data":"6a18ff8462e48fdd948b9a3c9d3556b78393b264c315dc137156c9c88559c223"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.083626 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7vh2m" podStartSLOduration=123.08360401 podStartE2EDuration="2m3.08360401s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:41.080214115 +0000 UTC m=+141.590825120" watchObservedRunningTime="2025-12-10 18:57:41.08360401 +0000 UTC m=+141.594215015" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.098400 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" event={"ID":"c5bd6649-abb0-4898-a73f-efceb62d2cb9","Type":"ContainerStarted","Data":"2d897701418eef8ce32adfb032b1a0b184b7fe03ec55abcb747335b88b3ebb7d"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.099058 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.101289 4828 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-w7n9w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.101347 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" podUID="c5bd6649-abb0-4898-a73f-efceb62d2cb9" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.102686 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-bwnph" event={"ID":"417523fb-3ff4-4747-a387-e9d2021d6006","Type":"ContainerStarted","Data":"0ee5066d29a5bb323f6f13faf5d4d1e0b6653374d4aed88cb75fdd84567177b2"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.121288 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" podStartSLOduration=123.121265854 podStartE2EDuration="2m3.121265854s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:41.120696605 +0000 UTC m=+141.631307610" watchObservedRunningTime="2025-12-10 18:57:41.121265854 +0000 UTC m=+141.631876859" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.125138 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" event={"ID":"9a8d7330-30c8-4a9c-ae2f-695b2c17deea","Type":"ContainerStarted","Data":"18fc18f747c979aefd932a223a03a9a3d881c9872526494dbfcceccb8bfc5cea"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.137137 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-22lss" event={"ID":"9c95e115-3ad7-46dc-b9cd-6a86f155512a","Type":"ContainerStarted","Data":"24ee364176becd49df8e89c4ad2037477f19e230574f5239c5f7f2559daf1362"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.146767 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.147828 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-7mhrd" podStartSLOduration=123.147810427 podStartE2EDuration="2m3.147810427s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:41.146097629 +0000 UTC m=+141.656708624" watchObservedRunningTime="2025-12-10 18:57:41.147810427 +0000 UTC m=+141.658421432" Dec 10 18:57:41 crc kubenswrapper[4828]: E1210 18:57:41.148216 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:41.648199743 +0000 UTC m=+142.158810748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.149477 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-x4r6k" event={"ID":"2cf82e3e-7ba7-4a3b-adf5-c635561b63d3","Type":"ContainerStarted","Data":"5a0c2b30dceffae3eea25e803c5567c834ec3daad1cbe51fc3e237a88256cc81"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.149531 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-x4r6k" event={"ID":"2cf82e3e-7ba7-4a3b-adf5-c635561b63d3","Type":"ContainerStarted","Data":"5b537ccd0cea1d183973d73dff1f49bd752d923ae030a299489aafd747193369"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.188534 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-bwnph" podStartSLOduration=6.188516321 podStartE2EDuration="6.188516321s" podCreationTimestamp="2025-12-10 18:57:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:41.182479522 +0000 UTC m=+141.693090527" watchObservedRunningTime="2025-12-10 18:57:41.188516321 +0000 UTC m=+141.699127326" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.201068 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wd865" event={"ID":"01750079-2ca2-4831-be15-a6d273d78680","Type":"ContainerStarted","Data":"2e7db66a490178ce2f1caf978b963e5eb8d404babc1abb937e807e318cdbf07b"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.230936 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" podStartSLOduration=122.230917692 podStartE2EDuration="2m2.230917692s" podCreationTimestamp="2025-12-10 18:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:41.213170643 +0000 UTC m=+141.723781648" watchObservedRunningTime="2025-12-10 18:57:41.230917692 +0000 UTC m=+141.741528697" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.241319 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncvkn" event={"ID":"09f4f7a0-5033-482d-9333-a2255ae9fb8c","Type":"ContainerStarted","Data":"a136c635fff0d46ad9a4af53e2a6e02c72c29879659015326ee399fbabadb57e"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.245376 4828 generic.go:334] "Generic (PLEG): container finished" podID="f59b0400-f54d-47cc-a58f-7a0399fa857d" containerID="204a98953b1640ac05748914fa5cc28cf182723cd1d96df7590d0643f0cab94d" exitCode=0 Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.245534 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" event={"ID":"f59b0400-f54d-47cc-a58f-7a0399fa857d","Type":"ContainerDied","Data":"204a98953b1640ac05748914fa5cc28cf182723cd1d96df7590d0643f0cab94d"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.248152 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:41 crc kubenswrapper[4828]: E1210 18:57:41.248436 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:41.748426367 +0000 UTC m=+142.259037372 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.251447 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf" event={"ID":"76b89fc6-2d65-415f-9680-f7a30d1217f1","Type":"ContainerStarted","Data":"7dafce669b5e4e5525940dad1412d3bd1c9129de9b7250ce59eee965459d93f1"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.252119 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.254125 4828 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-kmgkf container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.254282 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf" podUID="76b89fc6-2d65-415f-9680-f7a30d1217f1" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.256757 4828 generic.go:334] "Generic (PLEG): container finished" podID="9592c228-7f94-4923-818e-2d307bdf8d28" containerID="487a4558304f3047fc6481b7099c7e180a0fd25703dc9c22c5bcec83c29296fd" exitCode=0 Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.256887 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh" event={"ID":"9592c228-7f94-4923-818e-2d307bdf8d28","Type":"ContainerDied","Data":"487a4558304f3047fc6481b7099c7e180a0fd25703dc9c22c5bcec83c29296fd"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.268341 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" event={"ID":"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57","Type":"ContainerStarted","Data":"4a9e19d79cdb7c7a246bea9a129d3153a129b2e310c2a37bde6e515aa9bcd771"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.270826 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mm2tg" podStartSLOduration=122.270780502 podStartE2EDuration="2m2.270780502s" podCreationTimestamp="2025-12-10 18:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:41.233432593 +0000 UTC m=+141.744043598" watchObservedRunningTime="2025-12-10 18:57:41.270780502 +0000 UTC m=+141.781391507" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.281520 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" event={"ID":"48ee3baf-5f6c-48ed-b7f5-b82a4679f67f","Type":"ContainerStarted","Data":"253e22f9e843b3aefbf6ce1a096051f1147f26979667955c0d37142932cbd8e1"} Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.283168 4828 patch_prober.go:28] interesting pod/downloads-7954f5f757-s4zkd container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.283261 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-s4zkd" podUID="510cc49c-3ed7-4998-9d29-c50a7ab86337" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.19:8080/\": dial tcp 10.217.0.19:8080: connect: connection refused" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.288569 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ncvkn" podStartSLOduration=123.288538491 podStartE2EDuration="2m3.288538491s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:41.262217732 +0000 UTC m=+141.772828747" watchObservedRunningTime="2025-12-10 18:57:41.288538491 +0000 UTC m=+141.799149496" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.312263 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf" podStartSLOduration=122.312243308 podStartE2EDuration="2m2.312243308s" podCreationTimestamp="2025-12-10 18:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:41.31176691 +0000 UTC m=+141.822377915" watchObservedRunningTime="2025-12-10 18:57:41.312243308 +0000 UTC m=+141.822854303" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.350242 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:41 crc kubenswrapper[4828]: E1210 18:57:41.352558 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:41.852521554 +0000 UTC m=+142.363132559 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.379191 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" podStartSLOduration=122.379169859 podStartE2EDuration="2m2.379169859s" podCreationTimestamp="2025-12-10 18:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:41.354813822 +0000 UTC m=+141.865424837" watchObservedRunningTime="2025-12-10 18:57:41.379169859 +0000 UTC m=+141.889780864" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.425156 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" podStartSLOduration=123.425124558 podStartE2EDuration="2m3.425124558s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:41.400151651 +0000 UTC m=+141.910762656" watchObservedRunningTime="2025-12-10 18:57:41.425124558 +0000 UTC m=+141.935735563" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.453001 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:41 crc kubenswrapper[4828]: E1210 18:57:41.453822 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:41.953781685 +0000 UTC m=+142.464392680 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.453963 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-f92kp" podStartSLOduration=123.453939758 podStartE2EDuration="2m3.453939758s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:41.451213553 +0000 UTC m=+141.961824558" watchObservedRunningTime="2025-12-10 18:57:41.453939758 +0000 UTC m=+141.964550763" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.557944 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:41 crc kubenswrapper[4828]: E1210 18:57:41.558468 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:42.058450731 +0000 UTC m=+142.569061736 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.659926 4828 patch_prober.go:28] interesting pod/router-default-5444994796-qtdmw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:41 crc kubenswrapper[4828]: [-]has-synced failed: reason withheld Dec 10 18:57:41 crc kubenswrapper[4828]: [+]process-running ok Dec 10 18:57:41 crc kubenswrapper[4828]: healthz check failed Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.660260 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qtdmw" podUID="47cfb7d6-3723-4269-95ec-f6a512335387" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.660672 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:41 crc kubenswrapper[4828]: E1210 18:57:41.661103 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:42.161092025 +0000 UTC m=+142.671703030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.764323 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:41 crc kubenswrapper[4828]: E1210 18:57:41.764710 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:42.264693683 +0000 UTC m=+142.775304688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.867509 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:41 crc kubenswrapper[4828]: E1210 18:57:41.868155 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:42.36813938 +0000 UTC m=+142.878750385 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.968610 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:41 crc kubenswrapper[4828]: E1210 18:57:41.968755 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:42.46872864 +0000 UTC m=+142.979339645 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:41 crc kubenswrapper[4828]: I1210 18:57:41.968852 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:41 crc kubenswrapper[4828]: E1210 18:57:41.969162 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:42.469150547 +0000 UTC m=+142.979761552 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.070171 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:42 crc kubenswrapper[4828]: E1210 18:57:42.070359 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:42.570330316 +0000 UTC m=+143.080941321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.070479 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:42 crc kubenswrapper[4828]: E1210 18:57:42.070927 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:42.570920286 +0000 UTC m=+143.081531291 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.171898 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:42 crc kubenswrapper[4828]: E1210 18:57:42.172100 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:42.672042834 +0000 UTC m=+143.182653839 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.172428 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:42 crc kubenswrapper[4828]: E1210 18:57:42.172888 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:42.672869108 +0000 UTC m=+143.183480163 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.273232 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:42 crc kubenswrapper[4828]: E1210 18:57:42.273451 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:42.773417817 +0000 UTC m=+143.284028812 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.273507 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:42 crc kubenswrapper[4828]: E1210 18:57:42.273948 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:42.773935185 +0000 UTC m=+143.284546190 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.291994 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" event={"ID":"f59b0400-f54d-47cc-a58f-7a0399fa857d","Type":"ContainerStarted","Data":"1a83f0311ff028ba2a00401b32880a4ee402ba11de29d7bbf0c975519afa0e19"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.296284 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9jzvz" event={"ID":"2f1230e1-443a-43c5-bf94-40ef6201a476","Type":"ContainerStarted","Data":"9f116b7f07a57242569feb856812d4c3f3602409bea39583e8a3cae50a856680"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.296340 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9jzvz" event={"ID":"2f1230e1-443a-43c5-bf94-40ef6201a476","Type":"ContainerStarted","Data":"2e1552e4b945fec01df5e2ddc6b96c00b5abea1e8e6a7654d86d5b3c99e369f0"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.296428 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9jzvz" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.297639 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7x5g" event={"ID":"b0dc0bfb-c602-4e86-8e7f-9156696bdaf0","Type":"ContainerStarted","Data":"1808b38be37b078a2cf5da05878a896a211e66081f9ebcb43830ac73c4692304"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.303264 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vrmnl" event={"ID":"62e1b21d-36c2-4047-8913-8fc45e0401d7","Type":"ContainerStarted","Data":"bdcd88b25fcb0e3ae4f386de993693338d27d7e224d642bb9c98845732577cb2"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.303309 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-vrmnl" event={"ID":"62e1b21d-36c2-4047-8913-8fc45e0401d7","Type":"ContainerStarted","Data":"ebbaacc2dee13ca6a053bf4461e189e83e69a9cc109e8e51732c2eaea9dde7a9"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.307657 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mqr27" event={"ID":"c78748d2-622b-4cab-a462-793db246bc8e","Type":"ContainerStarted","Data":"e93a9898425c2eff0bcd0fcbea5b52b0f0b8f05cde498d38d092f3da11eb8d56"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.307721 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-mqr27" event={"ID":"c78748d2-622b-4cab-a462-793db246bc8e","Type":"ContainerStarted","Data":"caed34235c97f2600d457a20b5bc7adaed6528d53b9e7e0db4f7232f74e12d73"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.311070 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" event={"ID":"9a8d7330-30c8-4a9c-ae2f-695b2c17deea","Type":"ContainerStarted","Data":"c6dd6eaff1127d459c31eb2f1c230057a65687681c7a84fe7d0682d6d1bf085d"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.313760 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh" event={"ID":"9592c228-7f94-4923-818e-2d307bdf8d28","Type":"ContainerStarted","Data":"a4a33b76993b2f9ed20da97bc0524c86f44dfd3c17a7cc023fee004b9d3f19f0"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.313827 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.339608 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gdt2z" event={"ID":"109ecad0-ad50-4121-843e-319c8638705f","Type":"ContainerStarted","Data":"ef7dcc13966b9ac225bbe4d9670463c0f7bc83eb0a727cc776a2c98978bd5027"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.346334 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vfsz7" event={"ID":"de8e8906-1b66-4e41-8881-513f911a7407","Type":"ContainerStarted","Data":"e462bd3ad303c454d9a80b1af39bdf32fe8304073348e33694e5be01d8f36f96"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.359424 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-22lss" event={"ID":"9c95e115-3ad7-46dc-b9cd-6a86f155512a","Type":"ContainerStarted","Data":"fe5fe8bf83ed2c09d23750837e862ae1a3ad9aa54bf805de2ef3f635e1d98203"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.364740 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" event={"ID":"20d80266-84dc-4dee-83b6-3ccd9148fbca","Type":"ContainerStarted","Data":"19ee6e83dd0796c96ccc79d946b07dfd0afa1e4bbe6e91eecc44d95e8779a979"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.372080 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" podStartSLOduration=123.372058675 podStartE2EDuration="2m3.372058675s" podCreationTimestamp="2025-12-10 18:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:42.339460543 +0000 UTC m=+142.850071558" watchObservedRunningTime="2025-12-10 18:57:42.372058675 +0000 UTC m=+142.882669680" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.374741 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:42 crc kubenswrapper[4828]: E1210 18:57:42.374901 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:42.874882851 +0000 UTC m=+143.385493856 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.375347 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:42 crc kubenswrapper[4828]: E1210 18:57:42.377675 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:42.877664116 +0000 UTC m=+143.388275121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.387326 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" event={"ID":"e389c321-1fee-4503-8f27-ce2eb2dd76c3","Type":"ContainerStarted","Data":"193be46689f14130c97a864193eb734496a8a3169dd6dc0f29e7a7c11f3ce843"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.387630 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" event={"ID":"e389c321-1fee-4503-8f27-ce2eb2dd76c3","Type":"ContainerStarted","Data":"82c5bcec9c8d0f045a060dbf54c15180bb910ffa720be93b8a8058fe2f362c9a"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.391727 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hgg7d" event={"ID":"be16c9ea-01fe-4723-94f5-52a1f7a23cd9","Type":"ContainerStarted","Data":"20f9d3ef37a8286c47840e27ece1496beea057c6b18d78c49c30740cd7237447"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.391791 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-hgg7d" event={"ID":"be16c9ea-01fe-4723-94f5-52a1f7a23cd9","Type":"ContainerStarted","Data":"f633efdbe15c89b0584e9210f93433185d098ae8e1ef0c94d8af9ed1c82bdc93"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.392633 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-hgg7d" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.396669 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" event={"ID":"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57","Type":"ContainerStarted","Data":"56efab52e75e8cc9f2724828e354fd04d5ddca585f5c5ebd746b8de4d1a199a5"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.397971 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-x4r6k" event={"ID":"2cf82e3e-7ba7-4a3b-adf5-c635561b63d3","Type":"ContainerStarted","Data":"a6770c6533594954967fdbeb218d5a15ac66baeda6b340848c562b38471dfbac"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.399359 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wd865" event={"ID":"01750079-2ca2-4831-be15-a6d273d78680","Type":"ContainerStarted","Data":"1d64e6e97ff4e1cfcaf5d2c0661c285f3405efbd5490bee3276830730a732d1f"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.421895 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fv8kj" event={"ID":"493b4c81-95b7-44b8-a17b-f103d9602533","Type":"ContainerStarted","Data":"97291264486e595cb5e1a443370fe680259ace23fe9b99792252ca241152d0f3"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.424820 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-mqr27" podStartSLOduration=123.424785064 podStartE2EDuration="2m3.424785064s" podCreationTimestamp="2025-12-10 18:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:42.375145135 +0000 UTC m=+142.885756150" watchObservedRunningTime="2025-12-10 18:57:42.424785064 +0000 UTC m=+142.935396069" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.425980 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh" podStartSLOduration=124.425971044 podStartE2EDuration="2m4.425971044s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:42.423597795 +0000 UTC m=+142.934208800" watchObservedRunningTime="2025-12-10 18:57:42.425971044 +0000 UTC m=+142.936582039" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.450206 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf" event={"ID":"e8c0e685-9945-49df-818a-02daf5bb6099","Type":"ContainerStarted","Data":"4ccc9c51bc653f516898d639a7930a1545cc709d7ff86c1f151e8f111694180b"} Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.451513 4828 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-w7n9w container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.451575 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" podUID="c5bd6649-abb0-4898-a73f-efceb62d2cb9" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.24:8080/healthz\": dial tcp 10.217.0.24:8080: connect: connection refused" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.452657 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.452720 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.465036 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.470056 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-kmgkf" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.478379 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:42 crc kubenswrapper[4828]: E1210 18:57:42.480483 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:42.980470042 +0000 UTC m=+143.491081047 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.483518 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-r7x5g" podStartSLOduration=123.483495161 podStartE2EDuration="2m3.483495161s" podCreationTimestamp="2025-12-10 18:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:42.450120437 +0000 UTC m=+142.960731442" watchObservedRunningTime="2025-12-10 18:57:42.483495161 +0000 UTC m=+142.994106166" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.483844 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-p46hd" podStartSLOduration=124.483839517 podStartE2EDuration="2m4.483839517s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:42.482442454 +0000 UTC m=+142.993053449" watchObservedRunningTime="2025-12-10 18:57:42.483839517 +0000 UTC m=+142.994450512" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.526502 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9jzvz" podStartSLOduration=123.526485422 podStartE2EDuration="2m3.526485422s" podCreationTimestamp="2025-12-10 18:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:42.526213238 +0000 UTC m=+143.036824243" watchObservedRunningTime="2025-12-10 18:57:42.526485422 +0000 UTC m=+143.037096427" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.578667 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-vrmnl" podStartSLOduration=124.578649023 podStartE2EDuration="2m4.578649023s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:42.577930571 +0000 UTC m=+143.088541576" watchObservedRunningTime="2025-12-10 18:57:42.578649023 +0000 UTC m=+143.089260048" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.585307 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:42 crc kubenswrapper[4828]: E1210 18:57:42.585860 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:43.08582763 +0000 UTC m=+143.596438635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.589684 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.672707 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-jjrlf" podStartSLOduration=123.672689226 podStartE2EDuration="2m3.672689226s" podCreationTimestamp="2025-12-10 18:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:42.625680169 +0000 UTC m=+143.136291174" watchObservedRunningTime="2025-12-10 18:57:42.672689226 +0000 UTC m=+143.183300231" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.673513 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" podStartSLOduration=124.673507949 podStartE2EDuration="2m4.673507949s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:42.661039416 +0000 UTC m=+143.171650421" watchObservedRunningTime="2025-12-10 18:57:42.673507949 +0000 UTC m=+143.184118954" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.675406 4828 patch_prober.go:28] interesting pod/router-default-5444994796-qtdmw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:42 crc kubenswrapper[4828]: [-]has-synced failed: reason withheld Dec 10 18:57:42 crc kubenswrapper[4828]: [+]process-running ok Dec 10 18:57:42 crc kubenswrapper[4828]: healthz check failed Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.675465 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qtdmw" podUID="47cfb7d6-3723-4269-95ec-f6a512335387" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.697761 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:42 crc kubenswrapper[4828]: E1210 18:57:42.704854 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:43.2048327 +0000 UTC m=+143.715443705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.725848 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-22lss" podStartSLOduration=7.7258283720000005 podStartE2EDuration="7.725828372s" podCreationTimestamp="2025-12-10 18:57:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:42.723994492 +0000 UTC m=+143.234605517" watchObservedRunningTime="2025-12-10 18:57:42.725828372 +0000 UTC m=+143.236439377" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.726912 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" podStartSLOduration=124.726905009 podStartE2EDuration="2m4.726905009s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:42.697460679 +0000 UTC m=+143.208071684" watchObservedRunningTime="2025-12-10 18:57:42.726905009 +0000 UTC m=+143.237516014" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.734914 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.735348 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.760882 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" podStartSLOduration=124.760864403 podStartE2EDuration="2m4.760864403s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:42.760304734 +0000 UTC m=+143.270915749" watchObservedRunningTime="2025-12-10 18:57:42.760864403 +0000 UTC m=+143.271475398" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.807043 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:42 crc kubenswrapper[4828]: E1210 18:57:42.807569 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:43.307548184 +0000 UTC m=+143.818159259 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.848565 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-wd865" podStartSLOduration=123.848547572 podStartE2EDuration="2m3.848547572s" podCreationTimestamp="2025-12-10 18:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:42.846120933 +0000 UTC m=+143.356731938" watchObservedRunningTime="2025-12-10 18:57:42.848547572 +0000 UTC m=+143.359158577" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.855493 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.855554 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.859920 4828 patch_prober.go:28] interesting pod/apiserver-76f77b778f-8xs5l container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.15:8443/livez\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.859977 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" podUID="e389c321-1fee-4503-8f27-ce2eb2dd76c3" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.15:8443/livez\": dial tcp 10.217.0.15:8443: connect: connection refused" Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.908686 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:42 crc kubenswrapper[4828]: E1210 18:57:42.908955 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:43.408928267 +0000 UTC m=+143.919539272 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:42 crc kubenswrapper[4828]: I1210 18:57:42.964025 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fv8kj" podStartSLOduration=123.964003025 podStartE2EDuration="2m3.964003025s" podCreationTimestamp="2025-12-10 18:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:42.956100146 +0000 UTC m=+143.466711161" watchObservedRunningTime="2025-12-10 18:57:42.964003025 +0000 UTC m=+143.474614020" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.010379 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:43 crc kubenswrapper[4828]: E1210 18:57:43.010742 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:43.510730396 +0000 UTC m=+144.021341401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.035450 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gdt2z" podStartSLOduration=124.035431029 podStartE2EDuration="2m4.035431029s" podCreationTimestamp="2025-12-10 18:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:43.033709411 +0000 UTC m=+143.544320416" watchObservedRunningTime="2025-12-10 18:57:43.035431029 +0000 UTC m=+143.546042034" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.112078 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.112706 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-hgg7d" podStartSLOduration=8.112683267 podStartE2EDuration="8.112683267s" podCreationTimestamp="2025-12-10 18:57:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:43.10978254 +0000 UTC m=+143.620393545" watchObservedRunningTime="2025-12-10 18:57:43.112683267 +0000 UTC m=+143.623294272" Dec 10 18:57:43 crc kubenswrapper[4828]: E1210 18:57:43.112827 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:43.612807159 +0000 UTC m=+144.123418164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.193200 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-x4r6k" podStartSLOduration=125.1931854 podStartE2EDuration="2m5.1931854s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:43.192939646 +0000 UTC m=+143.703550651" watchObservedRunningTime="2025-12-10 18:57:43.1931854 +0000 UTC m=+143.703796405" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.215420 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:43 crc kubenswrapper[4828]: E1210 18:57:43.215981 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:43.715970771 +0000 UTC m=+144.226581776 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.322272 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:43 crc kubenswrapper[4828]: E1210 18:57:43.322662 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:43.82264718 +0000 UTC m=+144.333258185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.338898 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c8h8c" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.423460 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:43 crc kubenswrapper[4828]: E1210 18:57:43.424121 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:43.924108994 +0000 UTC m=+144.434719999 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.450172 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gs2g5"] Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.451281 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gs2g5" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.469547 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" event={"ID":"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57","Type":"ContainerStarted","Data":"82c9c7b28beaf8d71ebc9a6d9fe7f0be9b00f96ac7792311edf054467139c73a"} Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.475603 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.525517 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.525914 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxbj6\" (UniqueName: \"kubernetes.io/projected/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-kube-api-access-rxbj6\") pod \"certified-operators-gs2g5\" (UID: \"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc\") " pod="openshift-marketplace/certified-operators-gs2g5" Dec 10 18:57:43 crc kubenswrapper[4828]: E1210 18:57:43.526048 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:44.026018386 +0000 UTC m=+144.536629401 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.526229 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-utilities\") pod \"certified-operators-gs2g5\" (UID: \"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc\") " pod="openshift-marketplace/certified-operators-gs2g5" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.526683 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-catalog-content\") pod \"certified-operators-gs2g5\" (UID: \"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc\") " pod="openshift-marketplace/certified-operators-gs2g5" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.527867 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:43 crc kubenswrapper[4828]: E1210 18:57:43.528128 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:44.02811619 +0000 UTC m=+144.538727195 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.559018 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gs2g5"] Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.582603 4828 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.628919 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.629150 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxbj6\" (UniqueName: \"kubernetes.io/projected/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-kube-api-access-rxbj6\") pod \"certified-operators-gs2g5\" (UID: \"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc\") " pod="openshift-marketplace/certified-operators-gs2g5" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.629200 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-utilities\") pod \"certified-operators-gs2g5\" (UID: \"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc\") " pod="openshift-marketplace/certified-operators-gs2g5" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.629241 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-catalog-content\") pod \"certified-operators-gs2g5\" (UID: \"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc\") " pod="openshift-marketplace/certified-operators-gs2g5" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.629892 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-catalog-content\") pod \"certified-operators-gs2g5\" (UID: \"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc\") " pod="openshift-marketplace/certified-operators-gs2g5" Dec 10 18:57:43 crc kubenswrapper[4828]: E1210 18:57:43.629967 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:44.12995216 +0000 UTC m=+144.640563165 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.630407 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-utilities\") pod \"certified-operators-gs2g5\" (UID: \"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc\") " pod="openshift-marketplace/certified-operators-gs2g5" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.660783 4828 patch_prober.go:28] interesting pod/router-default-5444994796-qtdmw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:43 crc kubenswrapper[4828]: [-]has-synced failed: reason withheld Dec 10 18:57:43 crc kubenswrapper[4828]: [+]process-running ok Dec 10 18:57:43 crc kubenswrapper[4828]: healthz check failed Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.661540 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qtdmw" podUID="47cfb7d6-3723-4269-95ec-f6a512335387" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.666869 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7xlcc"] Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.671977 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7xlcc"] Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.672080 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7xlcc" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.674224 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.705329 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxbj6\" (UniqueName: \"kubernetes.io/projected/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-kube-api-access-rxbj6\") pod \"certified-operators-gs2g5\" (UID: \"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc\") " pod="openshift-marketplace/certified-operators-gs2g5" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.734303 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f95c086-c5d6-4ad2-97e6-79e652fa0040-catalog-content\") pod \"community-operators-7xlcc\" (UID: \"1f95c086-c5d6-4ad2-97e6-79e652fa0040\") " pod="openshift-marketplace/community-operators-7xlcc" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.734371 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f95c086-c5d6-4ad2-97e6-79e652fa0040-utilities\") pod \"community-operators-7xlcc\" (UID: \"1f95c086-c5d6-4ad2-97e6-79e652fa0040\") " pod="openshift-marketplace/community-operators-7xlcc" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.734413 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmntb\" (UniqueName: \"kubernetes.io/projected/1f95c086-c5d6-4ad2-97e6-79e652fa0040-kube-api-access-vmntb\") pod \"community-operators-7xlcc\" (UID: \"1f95c086-c5d6-4ad2-97e6-79e652fa0040\") " pod="openshift-marketplace/community-operators-7xlcc" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.734461 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:43 crc kubenswrapper[4828]: E1210 18:57:43.734904 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:44.234892371 +0000 UTC m=+144.745503376 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.747309 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.780919 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gs2g5" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.831312 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2vvgb"] Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.832296 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2vvgb" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.835695 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.836350 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f95c086-c5d6-4ad2-97e6-79e652fa0040-catalog-content\") pod \"community-operators-7xlcc\" (UID: \"1f95c086-c5d6-4ad2-97e6-79e652fa0040\") " pod="openshift-marketplace/community-operators-7xlcc" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.836413 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f95c086-c5d6-4ad2-97e6-79e652fa0040-utilities\") pod \"community-operators-7xlcc\" (UID: \"1f95c086-c5d6-4ad2-97e6-79e652fa0040\") " pod="openshift-marketplace/community-operators-7xlcc" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.836447 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmntb\" (UniqueName: \"kubernetes.io/projected/1f95c086-c5d6-4ad2-97e6-79e652fa0040-kube-api-access-vmntb\") pod \"community-operators-7xlcc\" (UID: \"1f95c086-c5d6-4ad2-97e6-79e652fa0040\") " pod="openshift-marketplace/community-operators-7xlcc" Dec 10 18:57:43 crc kubenswrapper[4828]: E1210 18:57:43.837111 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-10 18:57:44.337091017 +0000 UTC m=+144.847702022 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.837716 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f95c086-c5d6-4ad2-97e6-79e652fa0040-catalog-content\") pod \"community-operators-7xlcc\" (UID: \"1f95c086-c5d6-4ad2-97e6-79e652fa0040\") " pod="openshift-marketplace/community-operators-7xlcc" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.838069 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f95c086-c5d6-4ad2-97e6-79e652fa0040-utilities\") pod \"community-operators-7xlcc\" (UID: \"1f95c086-c5d6-4ad2-97e6-79e652fa0040\") " pod="openshift-marketplace/community-operators-7xlcc" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.846199 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2vvgb"] Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.879073 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmntb\" (UniqueName: \"kubernetes.io/projected/1f95c086-c5d6-4ad2-97e6-79e652fa0040-kube-api-access-vmntb\") pod \"community-operators-7xlcc\" (UID: \"1f95c086-c5d6-4ad2-97e6-79e652fa0040\") " pod="openshift-marketplace/community-operators-7xlcc" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.922860 4828 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-10T18:57:43.582631318Z","Handler":null,"Name":""} Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.938725 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43cd3400-d5e9-4279-9a85-051625ca4607-catalog-content\") pod \"certified-operators-2vvgb\" (UID: \"43cd3400-d5e9-4279-9a85-051625ca4607\") " pod="openshift-marketplace/certified-operators-2vvgb" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.938759 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43cd3400-d5e9-4279-9a85-051625ca4607-utilities\") pod \"certified-operators-2vvgb\" (UID: \"43cd3400-d5e9-4279-9a85-051625ca4607\") " pod="openshift-marketplace/certified-operators-2vvgb" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.938900 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.938948 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6z2dm\" (UniqueName: \"kubernetes.io/projected/43cd3400-d5e9-4279-9a85-051625ca4607-kube-api-access-6z2dm\") pod \"certified-operators-2vvgb\" (UID: \"43cd3400-d5e9-4279-9a85-051625ca4607\") " pod="openshift-marketplace/certified-operators-2vvgb" Dec 10 18:57:43 crc kubenswrapper[4828]: E1210 18:57:43.939393 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-10 18:57:44.439379834 +0000 UTC m=+144.949990839 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-c4wpt" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.962090 4828 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 10 18:57:43 crc kubenswrapper[4828]: I1210 18:57:43.962131 4828 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.032929 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7xlcc" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.041201 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.041415 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6z2dm\" (UniqueName: \"kubernetes.io/projected/43cd3400-d5e9-4279-9a85-051625ca4607-kube-api-access-6z2dm\") pod \"certified-operators-2vvgb\" (UID: \"43cd3400-d5e9-4279-9a85-051625ca4607\") " pod="openshift-marketplace/certified-operators-2vvgb" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.041465 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43cd3400-d5e9-4279-9a85-051625ca4607-catalog-content\") pod \"certified-operators-2vvgb\" (UID: \"43cd3400-d5e9-4279-9a85-051625ca4607\") " pod="openshift-marketplace/certified-operators-2vvgb" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.041480 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43cd3400-d5e9-4279-9a85-051625ca4607-utilities\") pod \"certified-operators-2vvgb\" (UID: \"43cd3400-d5e9-4279-9a85-051625ca4607\") " pod="openshift-marketplace/certified-operators-2vvgb" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.041888 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43cd3400-d5e9-4279-9a85-051625ca4607-utilities\") pod \"certified-operators-2vvgb\" (UID: \"43cd3400-d5e9-4279-9a85-051625ca4607\") " pod="openshift-marketplace/certified-operators-2vvgb" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.042292 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43cd3400-d5e9-4279-9a85-051625ca4607-catalog-content\") pod \"certified-operators-2vvgb\" (UID: \"43cd3400-d5e9-4279-9a85-051625ca4607\") " pod="openshift-marketplace/certified-operators-2vvgb" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.042500 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ll5gw"] Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.048940 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ll5gw" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.066213 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.074624 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ll5gw"] Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.079933 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6z2dm\" (UniqueName: \"kubernetes.io/projected/43cd3400-d5e9-4279-9a85-051625ca4607-kube-api-access-6z2dm\") pod \"certified-operators-2vvgb\" (UID: \"43cd3400-d5e9-4279-9a85-051625ca4607\") " pod="openshift-marketplace/certified-operators-2vvgb" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.143375 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.143455 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-catalog-content\") pod \"community-operators-ll5gw\" (UID: \"44b88622-40eb-4bcf-81a5-dab1f8b4c63b\") " pod="openshift-marketplace/community-operators-ll5gw" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.143524 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7gp8\" (UniqueName: \"kubernetes.io/projected/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-kube-api-access-w7gp8\") pod \"community-operators-ll5gw\" (UID: \"44b88622-40eb-4bcf-81a5-dab1f8b4c63b\") " pod="openshift-marketplace/community-operators-ll5gw" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.143555 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-utilities\") pod \"community-operators-ll5gw\" (UID: \"44b88622-40eb-4bcf-81a5-dab1f8b4c63b\") " pod="openshift-marketplace/community-operators-ll5gw" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.192028 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2vvgb" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.245966 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-catalog-content\") pod \"community-operators-ll5gw\" (UID: \"44b88622-40eb-4bcf-81a5-dab1f8b4c63b\") " pod="openshift-marketplace/community-operators-ll5gw" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.246036 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7gp8\" (UniqueName: \"kubernetes.io/projected/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-kube-api-access-w7gp8\") pod \"community-operators-ll5gw\" (UID: \"44b88622-40eb-4bcf-81a5-dab1f8b4c63b\") " pod="openshift-marketplace/community-operators-ll5gw" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.246093 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-utilities\") pod \"community-operators-ll5gw\" (UID: \"44b88622-40eb-4bcf-81a5-dab1f8b4c63b\") " pod="openshift-marketplace/community-operators-ll5gw" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.246953 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-utilities\") pod \"community-operators-ll5gw\" (UID: \"44b88622-40eb-4bcf-81a5-dab1f8b4c63b\") " pod="openshift-marketplace/community-operators-ll5gw" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.247184 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-catalog-content\") pod \"community-operators-ll5gw\" (UID: \"44b88622-40eb-4bcf-81a5-dab1f8b4c63b\") " pod="openshift-marketplace/community-operators-ll5gw" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.288640 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7gp8\" (UniqueName: \"kubernetes.io/projected/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-kube-api-access-w7gp8\") pod \"community-operators-ll5gw\" (UID: \"44b88622-40eb-4bcf-81a5-dab1f8b4c63b\") " pod="openshift-marketplace/community-operators-ll5gw" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.387675 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ll5gw" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.494160 4828 generic.go:334] "Generic (PLEG): container finished" podID="20d80266-84dc-4dee-83b6-3ccd9148fbca" containerID="19ee6e83dd0796c96ccc79d946b07dfd0afa1e4bbe6e91eecc44d95e8779a979" exitCode=0 Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.494437 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" event={"ID":"20d80266-84dc-4dee-83b6-3ccd9148fbca","Type":"ContainerDied","Data":"19ee6e83dd0796c96ccc79d946b07dfd0afa1e4bbe6e91eecc44d95e8779a979"} Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.510965 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-8smt2" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.519022 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gs2g5"] Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.617279 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7xlcc"] Dec 10 18:57:44 crc kubenswrapper[4828]: W1210 18:57:44.652988 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f95c086_c5d6_4ad2_97e6_79e652fa0040.slice/crio-3e9a991f46cc58755029cd47d9c6053f6d65187a531e9540fef390e2a5013f6b WatchSource:0}: Error finding container 3e9a991f46cc58755029cd47d9c6053f6d65187a531e9540fef390e2a5013f6b: Status 404 returned error can't find the container with id 3e9a991f46cc58755029cd47d9c6053f6d65187a531e9540fef390e2a5013f6b Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.686320 4828 patch_prober.go:28] interesting pod/router-default-5444994796-qtdmw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:44 crc kubenswrapper[4828]: [-]has-synced failed: reason withheld Dec 10 18:57:44 crc kubenswrapper[4828]: [+]process-running ok Dec 10 18:57:44 crc kubenswrapper[4828]: healthz check failed Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.686368 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qtdmw" podUID="47cfb7d6-3723-4269-95ec-f6a512335387" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.697499 4828 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.697547 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.777275 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2vvgb"] Dec 10 18:57:44 crc kubenswrapper[4828]: I1210 18:57:44.873513 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ll5gw"] Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.022744 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-c4wpt\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.060494 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.440455 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qs5qt"] Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.443719 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qs5qt" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.449918 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.457401 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qs5qt"] Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.484167 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9t2q\" (UniqueName: \"kubernetes.io/projected/b6084fc6-1311-4715-9d6d-01b42fec5f4b-kube-api-access-n9t2q\") pod \"redhat-marketplace-qs5qt\" (UID: \"b6084fc6-1311-4715-9d6d-01b42fec5f4b\") " pod="openshift-marketplace/redhat-marketplace-qs5qt" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.484234 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6084fc6-1311-4715-9d6d-01b42fec5f4b-catalog-content\") pod \"redhat-marketplace-qs5qt\" (UID: \"b6084fc6-1311-4715-9d6d-01b42fec5f4b\") " pod="openshift-marketplace/redhat-marketplace-qs5qt" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.484283 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6084fc6-1311-4715-9d6d-01b42fec5f4b-utilities\") pod \"redhat-marketplace-qs5qt\" (UID: \"b6084fc6-1311-4715-9d6d-01b42fec5f4b\") " pod="openshift-marketplace/redhat-marketplace-qs5qt" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.502165 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" event={"ID":"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57","Type":"ContainerStarted","Data":"023a9165c7e10e03a42e38ecaf027c179536f93ac4d00a5e217cee68a244293d"} Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.503664 4828 generic.go:334] "Generic (PLEG): container finished" podID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" containerID="d7b83a7de5179e88d2fd506f49860f038da6427203204ff64cdf56a978dc792e" exitCode=0 Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.503719 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ll5gw" event={"ID":"44b88622-40eb-4bcf-81a5-dab1f8b4c63b","Type":"ContainerDied","Data":"d7b83a7de5179e88d2fd506f49860f038da6427203204ff64cdf56a978dc792e"} Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.503735 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ll5gw" event={"ID":"44b88622-40eb-4bcf-81a5-dab1f8b4c63b","Type":"ContainerStarted","Data":"253088b4f1912303196e311ca62da10c9615c2eb2b30772f92278d42a79e2264"} Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.505257 4828 generic.go:334] "Generic (PLEG): container finished" podID="43cd3400-d5e9-4279-9a85-051625ca4607" containerID="442bd248c218c33c47b0aec9abca9ea73132a233d7290ac75c33dd4c60ee9ced" exitCode=0 Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.505342 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2vvgb" event={"ID":"43cd3400-d5e9-4279-9a85-051625ca4607","Type":"ContainerDied","Data":"442bd248c218c33c47b0aec9abca9ea73132a233d7290ac75c33dd4c60ee9ced"} Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.505366 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2vvgb" event={"ID":"43cd3400-d5e9-4279-9a85-051625ca4607","Type":"ContainerStarted","Data":"496dfddc50d3cc05bbad0965d627b7e72f2799b2fb422c04d246b4ea745cd1b5"} Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.505562 4828 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.507338 4828 generic.go:334] "Generic (PLEG): container finished" podID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" containerID="bbbaacb63de04ed7a04804f53dcde560ec5ac83aa5ab751bf45b94e6876c57cb" exitCode=0 Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.507391 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gs2g5" event={"ID":"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc","Type":"ContainerDied","Data":"bbbaacb63de04ed7a04804f53dcde560ec5ac83aa5ab751bf45b94e6876c57cb"} Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.507413 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gs2g5" event={"ID":"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc","Type":"ContainerStarted","Data":"07cf7a4a2d61c5484c036d16fb34f3779c57e1774cec637701262cdd09edaf90"} Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.509559 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f95c086-c5d6-4ad2-97e6-79e652fa0040" containerID="b036d88617a44b0d9fdb9d1570cb2a10a128f97f9fb9e308da23fb200a861ad7" exitCode=0 Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.509670 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xlcc" event={"ID":"1f95c086-c5d6-4ad2-97e6-79e652fa0040","Type":"ContainerDied","Data":"b036d88617a44b0d9fdb9d1570cb2a10a128f97f9fb9e308da23fb200a861ad7"} Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.509701 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xlcc" event={"ID":"1f95c086-c5d6-4ad2-97e6-79e652fa0040","Type":"ContainerStarted","Data":"3e9a991f46cc58755029cd47d9c6053f6d65187a531e9540fef390e2a5013f6b"} Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.585460 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6084fc6-1311-4715-9d6d-01b42fec5f4b-utilities\") pod \"redhat-marketplace-qs5qt\" (UID: \"b6084fc6-1311-4715-9d6d-01b42fec5f4b\") " pod="openshift-marketplace/redhat-marketplace-qs5qt" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.585928 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9t2q\" (UniqueName: \"kubernetes.io/projected/b6084fc6-1311-4715-9d6d-01b42fec5f4b-kube-api-access-n9t2q\") pod \"redhat-marketplace-qs5qt\" (UID: \"b6084fc6-1311-4715-9d6d-01b42fec5f4b\") " pod="openshift-marketplace/redhat-marketplace-qs5qt" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.586016 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6084fc6-1311-4715-9d6d-01b42fec5f4b-catalog-content\") pod \"redhat-marketplace-qs5qt\" (UID: \"b6084fc6-1311-4715-9d6d-01b42fec5f4b\") " pod="openshift-marketplace/redhat-marketplace-qs5qt" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.587130 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6084fc6-1311-4715-9d6d-01b42fec5f4b-utilities\") pod \"redhat-marketplace-qs5qt\" (UID: \"b6084fc6-1311-4715-9d6d-01b42fec5f4b\") " pod="openshift-marketplace/redhat-marketplace-qs5qt" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.589334 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6084fc6-1311-4715-9d6d-01b42fec5f4b-catalog-content\") pod \"redhat-marketplace-qs5qt\" (UID: \"b6084fc6-1311-4715-9d6d-01b42fec5f4b\") " pod="openshift-marketplace/redhat-marketplace-qs5qt" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.621842 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9t2q\" (UniqueName: \"kubernetes.io/projected/b6084fc6-1311-4715-9d6d-01b42fec5f4b-kube-api-access-n9t2q\") pod \"redhat-marketplace-qs5qt\" (UID: \"b6084fc6-1311-4715-9d6d-01b42fec5f4b\") " pod="openshift-marketplace/redhat-marketplace-qs5qt" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.640099 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c4wpt"] Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.661318 4828 patch_prober.go:28] interesting pod/router-default-5444994796-qtdmw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:45 crc kubenswrapper[4828]: [-]has-synced failed: reason withheld Dec 10 18:57:45 crc kubenswrapper[4828]: [+]process-running ok Dec 10 18:57:45 crc kubenswrapper[4828]: healthz check failed Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.661629 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qtdmw" podUID="47cfb7d6-3723-4269-95ec-f6a512335387" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.762002 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qs5qt" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.790461 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.797045 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.824282 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ftr4x"] Dec 10 18:57:45 crc kubenswrapper[4828]: E1210 18:57:45.824503 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20d80266-84dc-4dee-83b6-3ccd9148fbca" containerName="collect-profiles" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.824518 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="20d80266-84dc-4dee-83b6-3ccd9148fbca" containerName="collect-profiles" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.824603 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="20d80266-84dc-4dee-83b6-3ccd9148fbca" containerName="collect-profiles" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.825241 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ftr4x" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.835415 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ftr4x"] Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.877840 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.878589 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.880627 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.880917 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.882486 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.892221 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tnmp\" (UniqueName: \"kubernetes.io/projected/20d80266-84dc-4dee-83b6-3ccd9148fbca-kube-api-access-4tnmp\") pod \"20d80266-84dc-4dee-83b6-3ccd9148fbca\" (UID: \"20d80266-84dc-4dee-83b6-3ccd9148fbca\") " Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.892422 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20d80266-84dc-4dee-83b6-3ccd9148fbca-config-volume\") pod \"20d80266-84dc-4dee-83b6-3ccd9148fbca\" (UID: \"20d80266-84dc-4dee-83b6-3ccd9148fbca\") " Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.892492 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20d80266-84dc-4dee-83b6-3ccd9148fbca-secret-volume\") pod \"20d80266-84dc-4dee-83b6-3ccd9148fbca\" (UID: \"20d80266-84dc-4dee-83b6-3ccd9148fbca\") " Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.892731 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ef84f18-e802-4038-ba08-2b4eb948d803-catalog-content\") pod \"redhat-marketplace-ftr4x\" (UID: \"3ef84f18-e802-4038-ba08-2b4eb948d803\") " pod="openshift-marketplace/redhat-marketplace-ftr4x" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.892750 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ef84f18-e802-4038-ba08-2b4eb948d803-utilities\") pod \"redhat-marketplace-ftr4x\" (UID: \"3ef84f18-e802-4038-ba08-2b4eb948d803\") " pod="openshift-marketplace/redhat-marketplace-ftr4x" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.892827 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg6cp\" (UniqueName: \"kubernetes.io/projected/3ef84f18-e802-4038-ba08-2b4eb948d803-kube-api-access-wg6cp\") pod \"redhat-marketplace-ftr4x\" (UID: \"3ef84f18-e802-4038-ba08-2b4eb948d803\") " pod="openshift-marketplace/redhat-marketplace-ftr4x" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.893968 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20d80266-84dc-4dee-83b6-3ccd9148fbca-config-volume" (OuterVolumeSpecName: "config-volume") pod "20d80266-84dc-4dee-83b6-3ccd9148fbca" (UID: "20d80266-84dc-4dee-83b6-3ccd9148fbca"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.897701 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20d80266-84dc-4dee-83b6-3ccd9148fbca-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "20d80266-84dc-4dee-83b6-3ccd9148fbca" (UID: "20d80266-84dc-4dee-83b6-3ccd9148fbca"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.897863 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20d80266-84dc-4dee-83b6-3ccd9148fbca-kube-api-access-4tnmp" (OuterVolumeSpecName: "kube-api-access-4tnmp") pod "20d80266-84dc-4dee-83b6-3ccd9148fbca" (UID: "20d80266-84dc-4dee-83b6-3ccd9148fbca"). InnerVolumeSpecName "kube-api-access-4tnmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.967999 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qs5qt"] Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.994287 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d9e65b1c-141a-42b6-8bc6-0cc79293e76e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d9e65b1c-141a-42b6-8bc6-0cc79293e76e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.994337 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ef84f18-e802-4038-ba08-2b4eb948d803-catalog-content\") pod \"redhat-marketplace-ftr4x\" (UID: \"3ef84f18-e802-4038-ba08-2b4eb948d803\") " pod="openshift-marketplace/redhat-marketplace-ftr4x" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.994364 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ef84f18-e802-4038-ba08-2b4eb948d803-utilities\") pod \"redhat-marketplace-ftr4x\" (UID: \"3ef84f18-e802-4038-ba08-2b4eb948d803\") " pod="openshift-marketplace/redhat-marketplace-ftr4x" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.994447 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg6cp\" (UniqueName: \"kubernetes.io/projected/3ef84f18-e802-4038-ba08-2b4eb948d803-kube-api-access-wg6cp\") pod \"redhat-marketplace-ftr4x\" (UID: \"3ef84f18-e802-4038-ba08-2b4eb948d803\") " pod="openshift-marketplace/redhat-marketplace-ftr4x" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.994500 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d9e65b1c-141a-42b6-8bc6-0cc79293e76e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d9e65b1c-141a-42b6-8bc6-0cc79293e76e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.994554 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tnmp\" (UniqueName: \"kubernetes.io/projected/20d80266-84dc-4dee-83b6-3ccd9148fbca-kube-api-access-4tnmp\") on node \"crc\" DevicePath \"\"" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.994567 4828 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20d80266-84dc-4dee-83b6-3ccd9148fbca-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.994589 4828 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20d80266-84dc-4dee-83b6-3ccd9148fbca-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.995060 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ef84f18-e802-4038-ba08-2b4eb948d803-utilities\") pod \"redhat-marketplace-ftr4x\" (UID: \"3ef84f18-e802-4038-ba08-2b4eb948d803\") " pod="openshift-marketplace/redhat-marketplace-ftr4x" Dec 10 18:57:45 crc kubenswrapper[4828]: I1210 18:57:45.995072 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ef84f18-e802-4038-ba08-2b4eb948d803-catalog-content\") pod \"redhat-marketplace-ftr4x\" (UID: \"3ef84f18-e802-4038-ba08-2b4eb948d803\") " pod="openshift-marketplace/redhat-marketplace-ftr4x" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.013653 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg6cp\" (UniqueName: \"kubernetes.io/projected/3ef84f18-e802-4038-ba08-2b4eb948d803-kube-api-access-wg6cp\") pod \"redhat-marketplace-ftr4x\" (UID: \"3ef84f18-e802-4038-ba08-2b4eb948d803\") " pod="openshift-marketplace/redhat-marketplace-ftr4x" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.096068 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d9e65b1c-141a-42b6-8bc6-0cc79293e76e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d9e65b1c-141a-42b6-8bc6-0cc79293e76e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.096147 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d9e65b1c-141a-42b6-8bc6-0cc79293e76e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d9e65b1c-141a-42b6-8bc6-0cc79293e76e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.096330 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d9e65b1c-141a-42b6-8bc6-0cc79293e76e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d9e65b1c-141a-42b6-8bc6-0cc79293e76e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.113523 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d9e65b1c-141a-42b6-8bc6-0cc79293e76e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d9e65b1c-141a-42b6-8bc6-0cc79293e76e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.140897 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ftr4x" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.197741 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.350189 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ftr4x"] Dec 10 18:57:46 crc kubenswrapper[4828]: W1210 18:57:46.366151 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ef84f18_e802_4038_ba08_2b4eb948d803.slice/crio-c60c7f9b2d1ddd727493825a7f9e6a3f74d3ddf259f324b4f791740a2807dffe WatchSource:0}: Error finding container c60c7f9b2d1ddd727493825a7f9e6a3f74d3ddf259f324b4f791740a2807dffe: Status 404 returned error can't find the container with id c60c7f9b2d1ddd727493825a7f9e6a3f74d3ddf259f324b4f791740a2807dffe Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.424186 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 10 18:57:46 crc kubenswrapper[4828]: W1210 18:57:46.453343 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podd9e65b1c_141a_42b6_8bc6_0cc79293e76e.slice/crio-d43a6987e6cc3c476c7446e462eecacb947eb1e316c949d1000abe90fa436d11 WatchSource:0}: Error finding container d43a6987e6cc3c476c7446e462eecacb947eb1e316c949d1000abe90fa436d11: Status 404 returned error can't find the container with id d43a6987e6cc3c476c7446e462eecacb947eb1e316c949d1000abe90fa436d11 Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.515539 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" event={"ID":"1e72c425-11b5-4fd4-9f71-9df4d4489475","Type":"ContainerStarted","Data":"be827b35061551e27e2ea93f8ae82720cf53c1222786b1adcdb55d1302abee78"} Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.515838 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.515850 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" event={"ID":"1e72c425-11b5-4fd4-9f71-9df4d4489475","Type":"ContainerStarted","Data":"2db450e3fd742a7c15741e18ba9365b0f4caa86e1cbcb37f123da33ddd3de90d"} Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.517330 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.517321 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj" event={"ID":"20d80266-84dc-4dee-83b6-3ccd9148fbca","Type":"ContainerDied","Data":"9a7006f0f651b1b97ad91a37d2457735d5b58ace10fcdccb370f09464d891442"} Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.517448 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a7006f0f651b1b97ad91a37d2457735d5b58ace10fcdccb370f09464d891442" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.522286 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" event={"ID":"3c954830-5a1b-4bc4-ad5b-4b297f2cbc57","Type":"ContainerStarted","Data":"30c8b6f30d225fa076e6d20350430d9a712928ac5371b10210cee134caccd7a7"} Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.523790 4828 generic.go:334] "Generic (PLEG): container finished" podID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" containerID="92c81df57e70189d396ba6ff6646f0c857cb35e8e37710822d1eafb5702214a8" exitCode=0 Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.523870 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs5qt" event={"ID":"b6084fc6-1311-4715-9d6d-01b42fec5f4b","Type":"ContainerDied","Data":"92c81df57e70189d396ba6ff6646f0c857cb35e8e37710822d1eafb5702214a8"} Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.523896 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs5qt" event={"ID":"b6084fc6-1311-4715-9d6d-01b42fec5f4b","Type":"ContainerStarted","Data":"6f482adc183180d7a611457a1dca9793deb7270bb71879b9563741f65c1c9c39"} Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.525660 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftr4x" event={"ID":"3ef84f18-e802-4038-ba08-2b4eb948d803","Type":"ContainerStarted","Data":"c7959dcffd3e6f35062fb5118ec6fc089830b0634d9bdbc808f808ce9aa7d801"} Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.525692 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftr4x" event={"ID":"3ef84f18-e802-4038-ba08-2b4eb948d803","Type":"ContainerStarted","Data":"c60c7f9b2d1ddd727493825a7f9e6a3f74d3ddf259f324b4f791740a2807dffe"} Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.527062 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d9e65b1c-141a-42b6-8bc6-0cc79293e76e","Type":"ContainerStarted","Data":"d43a6987e6cc3c476c7446e462eecacb947eb1e316c949d1000abe90fa436d11"} Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.533285 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" podStartSLOduration=128.53326669 podStartE2EDuration="2m8.53326669s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:46.532612379 +0000 UTC m=+147.043223404" watchObservedRunningTime="2025-12-10 18:57:46.53326669 +0000 UTC m=+147.043877695" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.587041 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-z2f4h" podStartSLOduration=11.587024226 podStartE2EDuration="11.587024226s" podCreationTimestamp="2025-12-10 18:57:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:46.585321109 +0000 UTC m=+147.095932114" watchObservedRunningTime="2025-12-10 18:57:46.587024226 +0000 UTC m=+147.097635231" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.631766 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.659771 4828 patch_prober.go:28] interesting pod/router-default-5444994796-qtdmw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:46 crc kubenswrapper[4828]: [-]has-synced failed: reason withheld Dec 10 18:57:46 crc kubenswrapper[4828]: [+]process-running ok Dec 10 18:57:46 crc kubenswrapper[4828]: healthz check failed Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.659861 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qtdmw" podUID="47cfb7d6-3723-4269-95ec-f6a512335387" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.814921 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.814994 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.816225 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.830006 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zrqbh" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.835030 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.839193 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5d8xc"] Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.840324 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5d8xc" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.842001 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5d8xc"] Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.842996 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.917079 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e51cf27-f229-4c52-b839-c5f957e0fa99-utilities\") pod \"redhat-operators-5d8xc\" (UID: \"2e51cf27-f229-4c52-b839-c5f957e0fa99\") " pod="openshift-marketplace/redhat-operators-5d8xc" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.917122 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpl52\" (UniqueName: \"kubernetes.io/projected/2e51cf27-f229-4c52-b839-c5f957e0fa99-kube-api-access-qpl52\") pod \"redhat-operators-5d8xc\" (UID: \"2e51cf27-f229-4c52-b839-c5f957e0fa99\") " pod="openshift-marketplace/redhat-operators-5d8xc" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.917151 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e51cf27-f229-4c52-b839-c5f957e0fa99-catalog-content\") pod \"redhat-operators-5d8xc\" (UID: \"2e51cf27-f229-4c52-b839-c5f957e0fa99\") " pod="openshift-marketplace/redhat-operators-5d8xc" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.917186 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.917204 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.923553 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:46 crc kubenswrapper[4828]: I1210 18:57:46.931467 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.013958 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.018527 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e51cf27-f229-4c52-b839-c5f957e0fa99-utilities\") pod \"redhat-operators-5d8xc\" (UID: \"2e51cf27-f229-4c52-b839-c5f957e0fa99\") " pod="openshift-marketplace/redhat-operators-5d8xc" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.018578 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpl52\" (UniqueName: \"kubernetes.io/projected/2e51cf27-f229-4c52-b839-c5f957e0fa99-kube-api-access-qpl52\") pod \"redhat-operators-5d8xc\" (UID: \"2e51cf27-f229-4c52-b839-c5f957e0fa99\") " pod="openshift-marketplace/redhat-operators-5d8xc" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.018603 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e51cf27-f229-4c52-b839-c5f957e0fa99-catalog-content\") pod \"redhat-operators-5d8xc\" (UID: \"2e51cf27-f229-4c52-b839-c5f957e0fa99\") " pod="openshift-marketplace/redhat-operators-5d8xc" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.019004 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e51cf27-f229-4c52-b839-c5f957e0fa99-utilities\") pod \"redhat-operators-5d8xc\" (UID: \"2e51cf27-f229-4c52-b839-c5f957e0fa99\") " pod="openshift-marketplace/redhat-operators-5d8xc" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.019039 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e51cf27-f229-4c52-b839-c5f957e0fa99-catalog-content\") pod \"redhat-operators-5d8xc\" (UID: \"2e51cf27-f229-4c52-b839-c5f957e0fa99\") " pod="openshift-marketplace/redhat-operators-5d8xc" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.029566 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.041007 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.183584 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpl52\" (UniqueName: \"kubernetes.io/projected/2e51cf27-f229-4c52-b839-c5f957e0fa99-kube-api-access-qpl52\") pod \"redhat-operators-5d8xc\" (UID: \"2e51cf27-f229-4c52-b839-c5f957e0fa99\") " pod="openshift-marketplace/redhat-operators-5d8xc" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.226662 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kfgtf"] Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.227742 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kfgtf" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.234499 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kfgtf"] Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.322835 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e466f0bb-f597-46e1-b27a-a361322a6dcd-catalog-content\") pod \"redhat-operators-kfgtf\" (UID: \"e466f0bb-f597-46e1-b27a-a361322a6dcd\") " pod="openshift-marketplace/redhat-operators-kfgtf" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.325153 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e466f0bb-f597-46e1-b27a-a361322a6dcd-utilities\") pod \"redhat-operators-kfgtf\" (UID: \"e466f0bb-f597-46e1-b27a-a361322a6dcd\") " pod="openshift-marketplace/redhat-operators-kfgtf" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.325225 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg6qz\" (UniqueName: \"kubernetes.io/projected/e466f0bb-f597-46e1-b27a-a361322a6dcd-kube-api-access-hg6qz\") pod \"redhat-operators-kfgtf\" (UID: \"e466f0bb-f597-46e1-b27a-a361322a6dcd\") " pod="openshift-marketplace/redhat-operators-kfgtf" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.426854 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e466f0bb-f597-46e1-b27a-a361322a6dcd-catalog-content\") pod \"redhat-operators-kfgtf\" (UID: \"e466f0bb-f597-46e1-b27a-a361322a6dcd\") " pod="openshift-marketplace/redhat-operators-kfgtf" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.427669 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e466f0bb-f597-46e1-b27a-a361322a6dcd-utilities\") pod \"redhat-operators-kfgtf\" (UID: \"e466f0bb-f597-46e1-b27a-a361322a6dcd\") " pod="openshift-marketplace/redhat-operators-kfgtf" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.427701 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg6qz\" (UniqueName: \"kubernetes.io/projected/e466f0bb-f597-46e1-b27a-a361322a6dcd-kube-api-access-hg6qz\") pod \"redhat-operators-kfgtf\" (UID: \"e466f0bb-f597-46e1-b27a-a361322a6dcd\") " pod="openshift-marketplace/redhat-operators-kfgtf" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.427584 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e466f0bb-f597-46e1-b27a-a361322a6dcd-catalog-content\") pod \"redhat-operators-kfgtf\" (UID: \"e466f0bb-f597-46e1-b27a-a361322a6dcd\") " pod="openshift-marketplace/redhat-operators-kfgtf" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.428817 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e466f0bb-f597-46e1-b27a-a361322a6dcd-utilities\") pod \"redhat-operators-kfgtf\" (UID: \"e466f0bb-f597-46e1-b27a-a361322a6dcd\") " pod="openshift-marketplace/redhat-operators-kfgtf" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.449545 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg6qz\" (UniqueName: \"kubernetes.io/projected/e466f0bb-f597-46e1-b27a-a361322a6dcd-kube-api-access-hg6qz\") pod \"redhat-operators-kfgtf\" (UID: \"e466f0bb-f597-46e1-b27a-a361322a6dcd\") " pod="openshift-marketplace/redhat-operators-kfgtf" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.471480 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5d8xc" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.543370 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d9e65b1c-141a-42b6-8bc6-0cc79293e76e","Type":"ContainerStarted","Data":"e291e5ff3c19cd27419c4f31766e72b5346aa7c96ce3a27aaf64cb05316cec30"} Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.548555 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ca8fd6ea8fe172122efdd221a957ed870adb580f64ec14b3c0d7d2e2afa9ec4e"} Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.548697 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kfgtf" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.551098 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-s4zkd" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.552130 4828 generic.go:334] "Generic (PLEG): container finished" podID="3ef84f18-e802-4038-ba08-2b4eb948d803" containerID="c7959dcffd3e6f35062fb5118ec6fc089830b0634d9bdbc808f808ce9aa7d801" exitCode=0 Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.552188 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftr4x" event={"ID":"3ef84f18-e802-4038-ba08-2b4eb948d803","Type":"ContainerDied","Data":"c7959dcffd3e6f35062fb5118ec6fc089830b0634d9bdbc808f808ce9aa7d801"} Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.562663 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.56264109 podStartE2EDuration="2.56264109s" podCreationTimestamp="2025-12-10 18:57:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:47.560020417 +0000 UTC m=+148.070631422" watchObservedRunningTime="2025-12-10 18:57:47.56264109 +0000 UTC m=+148.073252105" Dec 10 18:57:47 crc kubenswrapper[4828]: W1210 18:57:47.617406 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-cc37b7e20c6d37fde128888423e287054d3a358f6e652c1adc130ac219f308d3 WatchSource:0}: Error finding container cc37b7e20c6d37fde128888423e287054d3a358f6e652c1adc130ac219f308d3: Status 404 returned error can't find the container with id cc37b7e20c6d37fde128888423e287054d3a358f6e652c1adc130ac219f308d3 Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.655648 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.658770 4828 patch_prober.go:28] interesting pod/router-default-5444994796-qtdmw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:47 crc kubenswrapper[4828]: [-]has-synced failed: reason withheld Dec 10 18:57:47 crc kubenswrapper[4828]: [+]process-running ok Dec 10 18:57:47 crc kubenswrapper[4828]: healthz check failed Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.658840 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qtdmw" podUID="47cfb7d6-3723-4269-95ec-f6a512335387" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:47 crc kubenswrapper[4828]: W1210 18:57:47.669363 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-9208b78f42e31939c706e95ff232abd28e91e0ed57a17c4842f6b80bb058384c WatchSource:0}: Error finding container 9208b78f42e31939c706e95ff232abd28e91e0ed57a17c4842f6b80bb058384c: Status 404 returned error can't find the container with id 9208b78f42e31939c706e95ff232abd28e91e0ed57a17c4842f6b80bb058384c Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.768115 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.768161 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.778484 4828 patch_prober.go:28] interesting pod/console-f9d7485db-m4fkj container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.778533 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-m4fkj" podUID="a2629f52-3b35-4707-8a24-53d7bbb70465" containerName="console" probeResult="failure" output="Get \"https://10.217.0.27:8443/health\": dial tcp 10.217.0.27:8443: connect: connection refused" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.807007 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.807061 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.860447 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:47 crc kubenswrapper[4828]: I1210 18:57:47.875450 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-8xs5l" Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.116600 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.117244 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.119864 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.122729 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.144509 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9af07058-3b8e-4815-ba12-6856c7fd8b6c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9af07058-3b8e-4815-ba12-6856c7fd8b6c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.144562 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9af07058-3b8e-4815-ba12-6856c7fd8b6c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9af07058-3b8e-4815-ba12-6856c7fd8b6c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.185531 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5d8xc"] Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.217262 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.246277 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9af07058-3b8e-4815-ba12-6856c7fd8b6c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9af07058-3b8e-4815-ba12-6856c7fd8b6c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.246322 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9af07058-3b8e-4815-ba12-6856c7fd8b6c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9af07058-3b8e-4815-ba12-6856c7fd8b6c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.246448 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9af07058-3b8e-4815-ba12-6856c7fd8b6c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9af07058-3b8e-4815-ba12-6856c7fd8b6c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.261157 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kfgtf"] Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.290776 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9af07058-3b8e-4815-ba12-6856c7fd8b6c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9af07058-3b8e-4815-ba12-6856c7fd8b6c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.443655 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.568446 4828 generic.go:334] "Generic (PLEG): container finished" podID="d9e65b1c-141a-42b6-8bc6-0cc79293e76e" containerID="e291e5ff3c19cd27419c4f31766e72b5346aa7c96ce3a27aaf64cb05316cec30" exitCode=0 Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.568545 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d9e65b1c-141a-42b6-8bc6-0cc79293e76e","Type":"ContainerDied","Data":"e291e5ff3c19cd27419c4f31766e72b5346aa7c96ce3a27aaf64cb05316cec30"} Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.586321 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7c5f505d86007d7539c5388a2859cd49b856d1eab2bfb8c1308dc84b3a3f55d9"} Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.586359 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"cc37b7e20c6d37fde128888423e287054d3a358f6e652c1adc130ac219f308d3"} Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.592094 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfgtf" event={"ID":"e466f0bb-f597-46e1-b27a-a361322a6dcd","Type":"ContainerStarted","Data":"4750b46b74dfac1ddaa41feb82e2e92bd3c7c970e69a1ff9cbda62d7f28b35bb"} Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.593222 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.611400 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"226af2bdc6553343563c0e2ab9f9ef5777b6fe1450cb3e0580485d3459b53b3c"} Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.611441 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9208b78f42e31939c706e95ff232abd28e91e0ed57a17c4842f6b80bb058384c"} Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.612872 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.617182 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2f776e288f5de252cf6bdb02170d4407e102f74807ed0fe19c8dffff8264c3cc"} Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.635164 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5d8xc" event={"ID":"2e51cf27-f229-4c52-b839-c5f957e0fa99","Type":"ContainerStarted","Data":"e3d1ba51ed316152866acef073e5d0d5ee7fbd13299905e3805dc065ffc2c0e6"} Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.660342 4828 patch_prober.go:28] interesting pod/router-default-5444994796-qtdmw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:48 crc kubenswrapper[4828]: [-]has-synced failed: reason withheld Dec 10 18:57:48 crc kubenswrapper[4828]: [+]process-running ok Dec 10 18:57:48 crc kubenswrapper[4828]: healthz check failed Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.660437 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qtdmw" podUID="47cfb7d6-3723-4269-95ec-f6a512335387" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:48 crc kubenswrapper[4828]: I1210 18:57:48.726753 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 10 18:57:48 crc kubenswrapper[4828]: W1210 18:57:48.780431 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod9af07058_3b8e_4815_ba12_6856c7fd8b6c.slice/crio-9d8ee7dbb1a900ee4ca186beedee98efc9a5d6bae4af5c90f48b5091c63e77ba WatchSource:0}: Error finding container 9d8ee7dbb1a900ee4ca186beedee98efc9a5d6bae4af5c90f48b5091c63e77ba: Status 404 returned error can't find the container with id 9d8ee7dbb1a900ee4ca186beedee98efc9a5d6bae4af5c90f48b5091c63e77ba Dec 10 18:57:49 crc kubenswrapper[4828]: I1210 18:57:49.653668 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9af07058-3b8e-4815-ba12-6856c7fd8b6c","Type":"ContainerStarted","Data":"ac4954edcda67a3b050fcdb51643d72eda0e8018b81a8c43e707a6632c99b30f"} Dec 10 18:57:49 crc kubenswrapper[4828]: I1210 18:57:49.654289 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9af07058-3b8e-4815-ba12-6856c7fd8b6c","Type":"ContainerStarted","Data":"9d8ee7dbb1a900ee4ca186beedee98efc9a5d6bae4af5c90f48b5091c63e77ba"} Dec 10 18:57:49 crc kubenswrapper[4828]: I1210 18:57:49.658165 4828 generic.go:334] "Generic (PLEG): container finished" podID="e466f0bb-f597-46e1-b27a-a361322a6dcd" containerID="32acb243cd59bea695a829412e491cd934342d0e3457c6e66993db1cc091c887" exitCode=0 Dec 10 18:57:49 crc kubenswrapper[4828]: I1210 18:57:49.658489 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfgtf" event={"ID":"e466f0bb-f597-46e1-b27a-a361322a6dcd","Type":"ContainerDied","Data":"32acb243cd59bea695a829412e491cd934342d0e3457c6e66993db1cc091c887"} Dec 10 18:57:49 crc kubenswrapper[4828]: I1210 18:57:49.659638 4828 patch_prober.go:28] interesting pod/router-default-5444994796-qtdmw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:49 crc kubenswrapper[4828]: [-]has-synced failed: reason withheld Dec 10 18:57:49 crc kubenswrapper[4828]: [+]process-running ok Dec 10 18:57:49 crc kubenswrapper[4828]: healthz check failed Dec 10 18:57:49 crc kubenswrapper[4828]: I1210 18:57:49.659677 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qtdmw" podUID="47cfb7d6-3723-4269-95ec-f6a512335387" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:49 crc kubenswrapper[4828]: I1210 18:57:49.664684 4828 generic.go:334] "Generic (PLEG): container finished" podID="2e51cf27-f229-4c52-b839-c5f957e0fa99" containerID="b9a195a31ee8a73c0f3b8191ce4e21ba46ebaff3eb3780edaa59f7dd4eb56473" exitCode=0 Dec 10 18:57:49 crc kubenswrapper[4828]: I1210 18:57:49.665788 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5d8xc" event={"ID":"2e51cf27-f229-4c52-b839-c5f957e0fa99","Type":"ContainerDied","Data":"b9a195a31ee8a73c0f3b8191ce4e21ba46ebaff3eb3780edaa59f7dd4eb56473"} Dec 10 18:57:50 crc kubenswrapper[4828]: I1210 18:57:50.002650 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:57:50 crc kubenswrapper[4828]: I1210 18:57:50.081766 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d9e65b1c-141a-42b6-8bc6-0cc79293e76e-kubelet-dir\") pod \"d9e65b1c-141a-42b6-8bc6-0cc79293e76e\" (UID: \"d9e65b1c-141a-42b6-8bc6-0cc79293e76e\") " Dec 10 18:57:50 crc kubenswrapper[4828]: I1210 18:57:50.081991 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d9e65b1c-141a-42b6-8bc6-0cc79293e76e-kube-api-access\") pod \"d9e65b1c-141a-42b6-8bc6-0cc79293e76e\" (UID: \"d9e65b1c-141a-42b6-8bc6-0cc79293e76e\") " Dec 10 18:57:50 crc kubenswrapper[4828]: I1210 18:57:50.081872 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d9e65b1c-141a-42b6-8bc6-0cc79293e76e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d9e65b1c-141a-42b6-8bc6-0cc79293e76e" (UID: "d9e65b1c-141a-42b6-8bc6-0cc79293e76e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:57:50 crc kubenswrapper[4828]: I1210 18:57:50.082456 4828 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d9e65b1c-141a-42b6-8bc6-0cc79293e76e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 18:57:50 crc kubenswrapper[4828]: I1210 18:57:50.089070 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9e65b1c-141a-42b6-8bc6-0cc79293e76e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d9e65b1c-141a-42b6-8bc6-0cc79293e76e" (UID: "d9e65b1c-141a-42b6-8bc6-0cc79293e76e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:57:50 crc kubenswrapper[4828]: I1210 18:57:50.183400 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d9e65b1c-141a-42b6-8bc6-0cc79293e76e-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 18:57:50 crc kubenswrapper[4828]: I1210 18:57:50.367442 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-hgg7d" Dec 10 18:57:50 crc kubenswrapper[4828]: I1210 18:57:50.662119 4828 patch_prober.go:28] interesting pod/router-default-5444994796-qtdmw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:50 crc kubenswrapper[4828]: [-]has-synced failed: reason withheld Dec 10 18:57:50 crc kubenswrapper[4828]: [+]process-running ok Dec 10 18:57:50 crc kubenswrapper[4828]: healthz check failed Dec 10 18:57:50 crc kubenswrapper[4828]: I1210 18:57:50.662175 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qtdmw" podUID="47cfb7d6-3723-4269-95ec-f6a512335387" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:50 crc kubenswrapper[4828]: I1210 18:57:50.678049 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 10 18:57:50 crc kubenswrapper[4828]: I1210 18:57:50.678343 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d9e65b1c-141a-42b6-8bc6-0cc79293e76e","Type":"ContainerDied","Data":"d43a6987e6cc3c476c7446e462eecacb947eb1e316c949d1000abe90fa436d11"} Dec 10 18:57:50 crc kubenswrapper[4828]: I1210 18:57:50.678374 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d43a6987e6cc3c476c7446e462eecacb947eb1e316c949d1000abe90fa436d11" Dec 10 18:57:50 crc kubenswrapper[4828]: I1210 18:57:50.698466 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.698438599 podStartE2EDuration="2.698438599s" podCreationTimestamp="2025-12-10 18:57:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:57:50.693491188 +0000 UTC m=+151.204102193" watchObservedRunningTime="2025-12-10 18:57:50.698438599 +0000 UTC m=+151.209049614" Dec 10 18:57:51 crc kubenswrapper[4828]: I1210 18:57:51.231287 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 18:57:51 crc kubenswrapper[4828]: I1210 18:57:51.231396 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 18:57:51 crc kubenswrapper[4828]: I1210 18:57:51.658417 4828 patch_prober.go:28] interesting pod/router-default-5444994796-qtdmw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:51 crc kubenswrapper[4828]: [-]has-synced failed: reason withheld Dec 10 18:57:51 crc kubenswrapper[4828]: [+]process-running ok Dec 10 18:57:51 crc kubenswrapper[4828]: healthz check failed Dec 10 18:57:51 crc kubenswrapper[4828]: I1210 18:57:51.658522 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qtdmw" podUID="47cfb7d6-3723-4269-95ec-f6a512335387" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:51 crc kubenswrapper[4828]: I1210 18:57:51.686596 4828 generic.go:334] "Generic (PLEG): container finished" podID="9af07058-3b8e-4815-ba12-6856c7fd8b6c" containerID="ac4954edcda67a3b050fcdb51643d72eda0e8018b81a8c43e707a6632c99b30f" exitCode=0 Dec 10 18:57:51 crc kubenswrapper[4828]: I1210 18:57:51.686635 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9af07058-3b8e-4815-ba12-6856c7fd8b6c","Type":"ContainerDied","Data":"ac4954edcda67a3b050fcdb51643d72eda0e8018b81a8c43e707a6632c99b30f"} Dec 10 18:57:52 crc kubenswrapper[4828]: I1210 18:57:52.658184 4828 patch_prober.go:28] interesting pod/router-default-5444994796-qtdmw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:52 crc kubenswrapper[4828]: [-]has-synced failed: reason withheld Dec 10 18:57:52 crc kubenswrapper[4828]: [+]process-running ok Dec 10 18:57:52 crc kubenswrapper[4828]: healthz check failed Dec 10 18:57:52 crc kubenswrapper[4828]: I1210 18:57:52.658487 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qtdmw" podUID="47cfb7d6-3723-4269-95ec-f6a512335387" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:52 crc kubenswrapper[4828]: I1210 18:57:52.931716 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:57:53 crc kubenswrapper[4828]: I1210 18:57:53.039243 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9af07058-3b8e-4815-ba12-6856c7fd8b6c-kube-api-access\") pod \"9af07058-3b8e-4815-ba12-6856c7fd8b6c\" (UID: \"9af07058-3b8e-4815-ba12-6856c7fd8b6c\") " Dec 10 18:57:53 crc kubenswrapper[4828]: I1210 18:57:53.039454 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9af07058-3b8e-4815-ba12-6856c7fd8b6c-kubelet-dir\") pod \"9af07058-3b8e-4815-ba12-6856c7fd8b6c\" (UID: \"9af07058-3b8e-4815-ba12-6856c7fd8b6c\") " Dec 10 18:57:53 crc kubenswrapper[4828]: I1210 18:57:53.039553 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9af07058-3b8e-4815-ba12-6856c7fd8b6c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9af07058-3b8e-4815-ba12-6856c7fd8b6c" (UID: "9af07058-3b8e-4815-ba12-6856c7fd8b6c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:57:53 crc kubenswrapper[4828]: I1210 18:57:53.040026 4828 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9af07058-3b8e-4815-ba12-6856c7fd8b6c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 18:57:53 crc kubenswrapper[4828]: I1210 18:57:53.044445 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9af07058-3b8e-4815-ba12-6856c7fd8b6c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9af07058-3b8e-4815-ba12-6856c7fd8b6c" (UID: "9af07058-3b8e-4815-ba12-6856c7fd8b6c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:57:53 crc kubenswrapper[4828]: I1210 18:57:53.141900 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9af07058-3b8e-4815-ba12-6856c7fd8b6c-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 18:57:53 crc kubenswrapper[4828]: I1210 18:57:53.659337 4828 patch_prober.go:28] interesting pod/router-default-5444994796-qtdmw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:53 crc kubenswrapper[4828]: [-]has-synced failed: reason withheld Dec 10 18:57:53 crc kubenswrapper[4828]: [+]process-running ok Dec 10 18:57:53 crc kubenswrapper[4828]: healthz check failed Dec 10 18:57:53 crc kubenswrapper[4828]: I1210 18:57:53.659715 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qtdmw" podUID="47cfb7d6-3723-4269-95ec-f6a512335387" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:53 crc kubenswrapper[4828]: I1210 18:57:53.701040 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9af07058-3b8e-4815-ba12-6856c7fd8b6c","Type":"ContainerDied","Data":"9d8ee7dbb1a900ee4ca186beedee98efc9a5d6bae4af5c90f48b5091c63e77ba"} Dec 10 18:57:53 crc kubenswrapper[4828]: I1210 18:57:53.701106 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d8ee7dbb1a900ee4ca186beedee98efc9a5d6bae4af5c90f48b5091c63e77ba" Dec 10 18:57:53 crc kubenswrapper[4828]: I1210 18:57:53.701111 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 10 18:57:54 crc kubenswrapper[4828]: I1210 18:57:54.657845 4828 patch_prober.go:28] interesting pod/router-default-5444994796-qtdmw container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 10 18:57:54 crc kubenswrapper[4828]: [+]has-synced ok Dec 10 18:57:54 crc kubenswrapper[4828]: [+]process-running ok Dec 10 18:57:54 crc kubenswrapper[4828]: healthz check failed Dec 10 18:57:54 crc kubenswrapper[4828]: I1210 18:57:54.657930 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-qtdmw" podUID="47cfb7d6-3723-4269-95ec-f6a512335387" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 10 18:57:55 crc kubenswrapper[4828]: I1210 18:57:55.658369 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:55 crc kubenswrapper[4828]: I1210 18:57:55.660646 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-qtdmw" Dec 10 18:57:58 crc kubenswrapper[4828]: I1210 18:57:58.277673 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:57:58 crc kubenswrapper[4828]: I1210 18:57:58.283076 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 18:58:01 crc kubenswrapper[4828]: I1210 18:58:01.063108 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs\") pod \"network-metrics-daemon-74lmx\" (UID: \"472ff1d2-0936-4ee0-b738-dc26e2c03559\") " pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:58:01 crc kubenswrapper[4828]: I1210 18:58:01.069019 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/472ff1d2-0936-4ee0-b738-dc26e2c03559-metrics-certs\") pod \"network-metrics-daemon-74lmx\" (UID: \"472ff1d2-0936-4ee0-b738-dc26e2c03559\") " pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:58:01 crc kubenswrapper[4828]: I1210 18:58:01.209734 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-74lmx" Dec 10 18:58:05 crc kubenswrapper[4828]: I1210 18:58:05.065231 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 18:58:17 crc kubenswrapper[4828]: I1210 18:58:17.047613 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 10 18:58:18 crc kubenswrapper[4828]: I1210 18:58:18.335071 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-9jzvz" Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.115697 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 10 18:58:21 crc kubenswrapper[4828]: E1210 18:58:21.116363 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9af07058-3b8e-4815-ba12-6856c7fd8b6c" containerName="pruner" Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.116401 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="9af07058-3b8e-4815-ba12-6856c7fd8b6c" containerName="pruner" Dec 10 18:58:21 crc kubenswrapper[4828]: E1210 18:58:21.116420 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9e65b1c-141a-42b6-8bc6-0cc79293e76e" containerName="pruner" Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.116428 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9e65b1c-141a-42b6-8bc6-0cc79293e76e" containerName="pruner" Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.116583 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9e65b1c-141a-42b6-8bc6-0cc79293e76e" containerName="pruner" Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.116597 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="9af07058-3b8e-4815-ba12-6856c7fd8b6c" containerName="pruner" Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.117000 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.123358 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.123387 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.127779 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.230827 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.230907 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.247155 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/67d86d48-a626-4d3f-9c11-8ac6830cc759-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"67d86d48-a626-4d3f-9c11-8ac6830cc759\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.247291 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/67d86d48-a626-4d3f-9c11-8ac6830cc759-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"67d86d48-a626-4d3f-9c11-8ac6830cc759\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.348888 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/67d86d48-a626-4d3f-9c11-8ac6830cc759-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"67d86d48-a626-4d3f-9c11-8ac6830cc759\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.348926 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/67d86d48-a626-4d3f-9c11-8ac6830cc759-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"67d86d48-a626-4d3f-9c11-8ac6830cc759\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.349070 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/67d86d48-a626-4d3f-9c11-8ac6830cc759-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"67d86d48-a626-4d3f-9c11-8ac6830cc759\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.370623 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/67d86d48-a626-4d3f-9c11-8ac6830cc759-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"67d86d48-a626-4d3f-9c11-8ac6830cc759\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:21 crc kubenswrapper[4828]: I1210 18:58:21.449572 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:24 crc kubenswrapper[4828]: E1210 18:58:24.771896 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 10 18:58:24 crc kubenswrapper[4828]: E1210 18:58:24.772470 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6z2dm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-2vvgb_openshift-marketplace(43cd3400-d5e9-4279-9a85-051625ca4607): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 18:58:24 crc kubenswrapper[4828]: E1210 18:58:24.773645 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-2vvgb" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" Dec 10 18:58:25 crc kubenswrapper[4828]: E1210 18:58:25.767463 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-2vvgb" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" Dec 10 18:58:25 crc kubenswrapper[4828]: E1210 18:58:25.847076 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 10 18:58:25 crc kubenswrapper[4828]: E1210 18:58:25.847209 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vmntb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-7xlcc_openshift-marketplace(1f95c086-c5d6-4ad2-97e6-79e652fa0040): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 18:58:25 crc kubenswrapper[4828]: E1210 18:58:25.848373 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-7xlcc" podUID="1f95c086-c5d6-4ad2-97e6-79e652fa0040" Dec 10 18:58:26 crc kubenswrapper[4828]: I1210 18:58:26.305732 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 10 18:58:26 crc kubenswrapper[4828]: I1210 18:58:26.306691 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:26 crc kubenswrapper[4828]: I1210 18:58:26.319408 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 10 18:58:26 crc kubenswrapper[4828]: I1210 18:58:26.417722 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/867406bf-a6e6-4644-a18b-4fa778cea1a5-kube-api-access\") pod \"installer-9-crc\" (UID: \"867406bf-a6e6-4644-a18b-4fa778cea1a5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:26 crc kubenswrapper[4828]: I1210 18:58:26.418100 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/867406bf-a6e6-4644-a18b-4fa778cea1a5-kubelet-dir\") pod \"installer-9-crc\" (UID: \"867406bf-a6e6-4644-a18b-4fa778cea1a5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:26 crc kubenswrapper[4828]: I1210 18:58:26.418170 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/867406bf-a6e6-4644-a18b-4fa778cea1a5-var-lock\") pod \"installer-9-crc\" (UID: \"867406bf-a6e6-4644-a18b-4fa778cea1a5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:26 crc kubenswrapper[4828]: I1210 18:58:26.518870 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/867406bf-a6e6-4644-a18b-4fa778cea1a5-var-lock\") pod \"installer-9-crc\" (UID: \"867406bf-a6e6-4644-a18b-4fa778cea1a5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:26 crc kubenswrapper[4828]: I1210 18:58:26.518947 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/867406bf-a6e6-4644-a18b-4fa778cea1a5-kube-api-access\") pod \"installer-9-crc\" (UID: \"867406bf-a6e6-4644-a18b-4fa778cea1a5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:26 crc kubenswrapper[4828]: I1210 18:58:26.518967 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/867406bf-a6e6-4644-a18b-4fa778cea1a5-kubelet-dir\") pod \"installer-9-crc\" (UID: \"867406bf-a6e6-4644-a18b-4fa778cea1a5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:26 crc kubenswrapper[4828]: I1210 18:58:26.519026 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/867406bf-a6e6-4644-a18b-4fa778cea1a5-var-lock\") pod \"installer-9-crc\" (UID: \"867406bf-a6e6-4644-a18b-4fa778cea1a5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:26 crc kubenswrapper[4828]: I1210 18:58:26.519059 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/867406bf-a6e6-4644-a18b-4fa778cea1a5-kubelet-dir\") pod \"installer-9-crc\" (UID: \"867406bf-a6e6-4644-a18b-4fa778cea1a5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:26 crc kubenswrapper[4828]: I1210 18:58:26.539998 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/867406bf-a6e6-4644-a18b-4fa778cea1a5-kube-api-access\") pod \"installer-9-crc\" (UID: \"867406bf-a6e6-4644-a18b-4fa778cea1a5\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:26 crc kubenswrapper[4828]: I1210 18:58:26.628993 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:58:27 crc kubenswrapper[4828]: E1210 18:58:27.688616 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-7xlcc" podUID="1f95c086-c5d6-4ad2-97e6-79e652fa0040" Dec 10 18:58:27 crc kubenswrapper[4828]: E1210 18:58:27.774257 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 10 18:58:27 crc kubenswrapper[4828]: E1210 18:58:27.774489 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wg6cp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-ftr4x_openshift-marketplace(3ef84f18-e802-4038-ba08-2b4eb948d803): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 18:58:27 crc kubenswrapper[4828]: E1210 18:58:27.775647 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-ftr4x" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" Dec 10 18:58:27 crc kubenswrapper[4828]: E1210 18:58:27.789883 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 10 18:58:27 crc kubenswrapper[4828]: E1210 18:58:27.790081 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rxbj6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-gs2g5_openshift-marketplace(0dbff95a-c195-4a72-bbfe-0a31a3feb8cc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 18:58:27 crc kubenswrapper[4828]: E1210 18:58:27.795303 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-gs2g5" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" Dec 10 18:58:28 crc kubenswrapper[4828]: I1210 18:58:28.320758 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hgsxd"] Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.739648 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-gs2g5" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.739958 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ftr4x" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.789401 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.789873 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qpl52,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-5d8xc_openshift-marketplace(2e51cf27-f229-4c52-b839-c5f957e0fa99): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.791267 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-5d8xc" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.817463 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.818301 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hg6qz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-kfgtf_openshift-marketplace(e466f0bb-f597-46e1-b27a-a361322a6dcd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.820926 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-kfgtf" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.832840 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.833035 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n9t2q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-qs5qt_openshift-marketplace(b6084fc6-1311-4715-9d6d-01b42fec5f4b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.834564 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-qs5qt" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.851733 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.851921 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w7gp8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-ll5gw_openshift-marketplace(44b88622-40eb-4bcf-81a5-dab1f8b4c63b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.853716 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-ll5gw" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.926292 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-kfgtf" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.926711 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-5d8xc" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.926832 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-qs5qt" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" Dec 10 18:58:30 crc kubenswrapper[4828]: E1210 18:58:30.926925 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-ll5gw" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" Dec 10 18:58:31 crc kubenswrapper[4828]: I1210 18:58:31.153716 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 10 18:58:31 crc kubenswrapper[4828]: I1210 18:58:31.223752 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-74lmx"] Dec 10 18:58:31 crc kubenswrapper[4828]: W1210 18:58:31.227088 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod472ff1d2_0936_4ee0_b738_dc26e2c03559.slice/crio-8b819bba92fc231e9b4086f1963eee1605d67e8178e1be78ca1323b31259f39d WatchSource:0}: Error finding container 8b819bba92fc231e9b4086f1963eee1605d67e8178e1be78ca1323b31259f39d: Status 404 returned error can't find the container with id 8b819bba92fc231e9b4086f1963eee1605d67e8178e1be78ca1323b31259f39d Dec 10 18:58:31 crc kubenswrapper[4828]: I1210 18:58:31.301352 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 10 18:58:31 crc kubenswrapper[4828]: I1210 18:58:31.929196 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"867406bf-a6e6-4644-a18b-4fa778cea1a5","Type":"ContainerStarted","Data":"960167c72cf91fc50de7ec6b94de36dead12bbc4fe9ec62e1e0facb867ddd855"} Dec 10 18:58:31 crc kubenswrapper[4828]: I1210 18:58:31.930138 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"67d86d48-a626-4d3f-9c11-8ac6830cc759","Type":"ContainerStarted","Data":"c1eab7e87ddbe198f768df9ffff6f9fc83388517792f2ee31ba7df9eca9a2e14"} Dec 10 18:58:31 crc kubenswrapper[4828]: I1210 18:58:31.930846 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-74lmx" event={"ID":"472ff1d2-0936-4ee0-b738-dc26e2c03559","Type":"ContainerStarted","Data":"8b819bba92fc231e9b4086f1963eee1605d67e8178e1be78ca1323b31259f39d"} Dec 10 18:58:33 crc kubenswrapper[4828]: I1210 18:58:33.941628 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"67d86d48-a626-4d3f-9c11-8ac6830cc759","Type":"ContainerStarted","Data":"5ece9b032ce0f915b0e817c43f89953b554e6cd38f12c0de0d1d09c6b73d0361"} Dec 10 18:58:33 crc kubenswrapper[4828]: I1210 18:58:33.945356 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-74lmx" event={"ID":"472ff1d2-0936-4ee0-b738-dc26e2c03559","Type":"ContainerStarted","Data":"1dcb879ed9289c17338c5cbf00383058051ec1c3b2a1875e86c6ad1d1d4a9ec6"} Dec 10 18:58:33 crc kubenswrapper[4828]: I1210 18:58:33.945398 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-74lmx" event={"ID":"472ff1d2-0936-4ee0-b738-dc26e2c03559","Type":"ContainerStarted","Data":"7df97ba585d0fa0bf07826b28925a0acc84f278293b90ca4e7d1c9e3049afd51"} Dec 10 18:58:33 crc kubenswrapper[4828]: I1210 18:58:33.949069 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"867406bf-a6e6-4644-a18b-4fa778cea1a5","Type":"ContainerStarted","Data":"7e0bf15da9681106df3d2f8e11a52c960ea641077bfc3c112b437708170c197e"} Dec 10 18:58:33 crc kubenswrapper[4828]: I1210 18:58:33.959974 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=12.959956163 podStartE2EDuration="12.959956163s" podCreationTimestamp="2025-12-10 18:58:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:58:33.957682836 +0000 UTC m=+194.468293861" watchObservedRunningTime="2025-12-10 18:58:33.959956163 +0000 UTC m=+194.470567168" Dec 10 18:58:33 crc kubenswrapper[4828]: I1210 18:58:33.980909 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=7.980889274 podStartE2EDuration="7.980889274s" podCreationTimestamp="2025-12-10 18:58:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:58:33.976810988 +0000 UTC m=+194.487422013" watchObservedRunningTime="2025-12-10 18:58:33.980889274 +0000 UTC m=+194.491500279" Dec 10 18:58:34 crc kubenswrapper[4828]: I1210 18:58:34.955474 4828 generic.go:334] "Generic (PLEG): container finished" podID="67d86d48-a626-4d3f-9c11-8ac6830cc759" containerID="5ece9b032ce0f915b0e817c43f89953b554e6cd38f12c0de0d1d09c6b73d0361" exitCode=0 Dec 10 18:58:34 crc kubenswrapper[4828]: I1210 18:58:34.955551 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"67d86d48-a626-4d3f-9c11-8ac6830cc759","Type":"ContainerDied","Data":"5ece9b032ce0f915b0e817c43f89953b554e6cd38f12c0de0d1d09c6b73d0361"} Dec 10 18:58:34 crc kubenswrapper[4828]: I1210 18:58:34.970989 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-74lmx" podStartSLOduration=176.970969574 podStartE2EDuration="2m56.970969574s" podCreationTimestamp="2025-12-10 18:55:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:58:33.996175013 +0000 UTC m=+194.506786018" watchObservedRunningTime="2025-12-10 18:58:34.970969574 +0000 UTC m=+195.481580579" Dec 10 18:58:36 crc kubenswrapper[4828]: I1210 18:58:36.160834 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:36 crc kubenswrapper[4828]: I1210 18:58:36.254922 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/67d86d48-a626-4d3f-9c11-8ac6830cc759-kubelet-dir\") pod \"67d86d48-a626-4d3f-9c11-8ac6830cc759\" (UID: \"67d86d48-a626-4d3f-9c11-8ac6830cc759\") " Dec 10 18:58:36 crc kubenswrapper[4828]: I1210 18:58:36.254997 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/67d86d48-a626-4d3f-9c11-8ac6830cc759-kube-api-access\") pod \"67d86d48-a626-4d3f-9c11-8ac6830cc759\" (UID: \"67d86d48-a626-4d3f-9c11-8ac6830cc759\") " Dec 10 18:58:36 crc kubenswrapper[4828]: I1210 18:58:36.255025 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67d86d48-a626-4d3f-9c11-8ac6830cc759-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "67d86d48-a626-4d3f-9c11-8ac6830cc759" (UID: "67d86d48-a626-4d3f-9c11-8ac6830cc759"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:58:36 crc kubenswrapper[4828]: I1210 18:58:36.255376 4828 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/67d86d48-a626-4d3f-9c11-8ac6830cc759-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:36 crc kubenswrapper[4828]: I1210 18:58:36.268910 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d86d48-a626-4d3f-9c11-8ac6830cc759-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "67d86d48-a626-4d3f-9c11-8ac6830cc759" (UID: "67d86d48-a626-4d3f-9c11-8ac6830cc759"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:58:36 crc kubenswrapper[4828]: I1210 18:58:36.356324 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/67d86d48-a626-4d3f-9c11-8ac6830cc759-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:36 crc kubenswrapper[4828]: I1210 18:58:36.966706 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"67d86d48-a626-4d3f-9c11-8ac6830cc759","Type":"ContainerDied","Data":"c1eab7e87ddbe198f768df9ffff6f9fc83388517792f2ee31ba7df9eca9a2e14"} Dec 10 18:58:36 crc kubenswrapper[4828]: I1210 18:58:36.966766 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 10 18:58:36 crc kubenswrapper[4828]: I1210 18:58:36.966774 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1eab7e87ddbe198f768df9ffff6f9fc83388517792f2ee31ba7df9eca9a2e14" Dec 10 18:58:41 crc kubenswrapper[4828]: I1210 18:58:41.625887 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n9pnl"] Dec 10 18:58:41 crc kubenswrapper[4828]: I1210 18:58:41.626376 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" podUID="5f498d61-1987-48a1-82d2-14ea6db95c84" containerName="controller-manager" containerID="cri-o://3c38d62467fe683e8ac5876857183ae6676c9a18fa4b9c7c745e83472657f49a" gracePeriod=30 Dec 10 18:58:41 crc kubenswrapper[4828]: I1210 18:58:41.720929 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2"] Dec 10 18:58:41 crc kubenswrapper[4828]: I1210 18:58:41.721126 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" podUID="e9d9df65-e6cc-404d-8cf9-f50c566d5595" containerName="route-controller-manager" containerID="cri-o://23203906e6c8e9ab5602b14502fb40b91bdc27164c2394aa378d570566a2f1f8" gracePeriod=30 Dec 10 18:58:44 crc kubenswrapper[4828]: I1210 18:58:44.001972 4828 generic.go:334] "Generic (PLEG): container finished" podID="5f498d61-1987-48a1-82d2-14ea6db95c84" containerID="3c38d62467fe683e8ac5876857183ae6676c9a18fa4b9c7c745e83472657f49a" exitCode=0 Dec 10 18:58:44 crc kubenswrapper[4828]: I1210 18:58:44.002017 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" event={"ID":"5f498d61-1987-48a1-82d2-14ea6db95c84","Type":"ContainerDied","Data":"3c38d62467fe683e8ac5876857183ae6676c9a18fa4b9c7c745e83472657f49a"} Dec 10 18:58:45 crc kubenswrapper[4828]: I1210 18:58:45.013050 4828 generic.go:334] "Generic (PLEG): container finished" podID="e9d9df65-e6cc-404d-8cf9-f50c566d5595" containerID="23203906e6c8e9ab5602b14502fb40b91bdc27164c2394aa378d570566a2f1f8" exitCode=0 Dec 10 18:58:45 crc kubenswrapper[4828]: I1210 18:58:45.013101 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" event={"ID":"e9d9df65-e6cc-404d-8cf9-f50c566d5595","Type":"ContainerDied","Data":"23203906e6c8e9ab5602b14502fb40b91bdc27164c2394aa378d570566a2f1f8"} Dec 10 18:58:47 crc kubenswrapper[4828]: I1210 18:58:47.045950 4828 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-j8vl2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Dec 10 18:58:47 crc kubenswrapper[4828]: I1210 18:58:47.046268 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" podUID="e9d9df65-e6cc-404d-8cf9-f50c566d5595" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Dec 10 18:58:47 crc kubenswrapper[4828]: I1210 18:58:47.758633 4828 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-n9pnl container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 10 18:58:47 crc kubenswrapper[4828]: I1210 18:58:47.758695 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" podUID="5f498d61-1987-48a1-82d2-14ea6db95c84" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.158745 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.164542 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.193224 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv"] Dec 10 18:58:51 crc kubenswrapper[4828]: E1210 18:58:51.193533 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f498d61-1987-48a1-82d2-14ea6db95c84" containerName="controller-manager" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.193552 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f498d61-1987-48a1-82d2-14ea6db95c84" containerName="controller-manager" Dec 10 18:58:51 crc kubenswrapper[4828]: E1210 18:58:51.193567 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d86d48-a626-4d3f-9c11-8ac6830cc759" containerName="pruner" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.193578 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d86d48-a626-4d3f-9c11-8ac6830cc759" containerName="pruner" Dec 10 18:58:51 crc kubenswrapper[4828]: E1210 18:58:51.193603 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9d9df65-e6cc-404d-8cf9-f50c566d5595" containerName="route-controller-manager" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.193616 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9d9df65-e6cc-404d-8cf9-f50c566d5595" containerName="route-controller-manager" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.193767 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d86d48-a626-4d3f-9c11-8ac6830cc759" containerName="pruner" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.193782 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f498d61-1987-48a1-82d2-14ea6db95c84" containerName="controller-manager" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.193817 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9d9df65-e6cc-404d-8cf9-f50c566d5595" containerName="route-controller-manager" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.195185 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.202718 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv"] Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.231237 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.231316 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.231366 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.232150 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.232268 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71" gracePeriod=600 Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.249911 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d9df65-e6cc-404d-8cf9-f50c566d5595-serving-cert\") pod \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\" (UID: \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\") " Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.249946 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fq5cn\" (UniqueName: \"kubernetes.io/projected/5f498d61-1987-48a1-82d2-14ea6db95c84-kube-api-access-fq5cn\") pod \"5f498d61-1987-48a1-82d2-14ea6db95c84\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.250018 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d9df65-e6cc-404d-8cf9-f50c566d5595-config\") pod \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\" (UID: \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\") " Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.250071 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9d9df65-e6cc-404d-8cf9-f50c566d5595-client-ca\") pod \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\" (UID: \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\") " Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.250089 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-proxy-ca-bundles\") pod \"5f498d61-1987-48a1-82d2-14ea6db95c84\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.250121 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nwbk\" (UniqueName: \"kubernetes.io/projected/e9d9df65-e6cc-404d-8cf9-f50c566d5595-kube-api-access-4nwbk\") pod \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\" (UID: \"e9d9df65-e6cc-404d-8cf9-f50c566d5595\") " Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.250141 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f498d61-1987-48a1-82d2-14ea6db95c84-serving-cert\") pod \"5f498d61-1987-48a1-82d2-14ea6db95c84\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.250184 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-client-ca\") pod \"5f498d61-1987-48a1-82d2-14ea6db95c84\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.250204 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-config\") pod \"5f498d61-1987-48a1-82d2-14ea6db95c84\" (UID: \"5f498d61-1987-48a1-82d2-14ea6db95c84\") " Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.251650 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "5f498d61-1987-48a1-82d2-14ea6db95c84" (UID: "5f498d61-1987-48a1-82d2-14ea6db95c84"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.251788 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-config" (OuterVolumeSpecName: "config") pod "5f498d61-1987-48a1-82d2-14ea6db95c84" (UID: "5f498d61-1987-48a1-82d2-14ea6db95c84"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.251938 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9d9df65-e6cc-404d-8cf9-f50c566d5595-client-ca" (OuterVolumeSpecName: "client-ca") pod "e9d9df65-e6cc-404d-8cf9-f50c566d5595" (UID: "e9d9df65-e6cc-404d-8cf9-f50c566d5595"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.251956 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9d9df65-e6cc-404d-8cf9-f50c566d5595-config" (OuterVolumeSpecName: "config") pod "e9d9df65-e6cc-404d-8cf9-f50c566d5595" (UID: "e9d9df65-e6cc-404d-8cf9-f50c566d5595"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.252416 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-client-ca" (OuterVolumeSpecName: "client-ca") pod "5f498d61-1987-48a1-82d2-14ea6db95c84" (UID: "5f498d61-1987-48a1-82d2-14ea6db95c84"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.257703 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9d9df65-e6cc-404d-8cf9-f50c566d5595-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e9d9df65-e6cc-404d-8cf9-f50c566d5595" (UID: "e9d9df65-e6cc-404d-8cf9-f50c566d5595"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.258301 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9d9df65-e6cc-404d-8cf9-f50c566d5595-kube-api-access-4nwbk" (OuterVolumeSpecName: "kube-api-access-4nwbk") pod "e9d9df65-e6cc-404d-8cf9-f50c566d5595" (UID: "e9d9df65-e6cc-404d-8cf9-f50c566d5595"). InnerVolumeSpecName "kube-api-access-4nwbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.264545 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f498d61-1987-48a1-82d2-14ea6db95c84-kube-api-access-fq5cn" (OuterVolumeSpecName: "kube-api-access-fq5cn") pod "5f498d61-1987-48a1-82d2-14ea6db95c84" (UID: "5f498d61-1987-48a1-82d2-14ea6db95c84"). InnerVolumeSpecName "kube-api-access-fq5cn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.268225 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f498d61-1987-48a1-82d2-14ea6db95c84-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5f498d61-1987-48a1-82d2-14ea6db95c84" (UID: "5f498d61-1987-48a1-82d2-14ea6db95c84"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.350980 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zph6h\" (UniqueName: \"kubernetes.io/projected/85621df9-138d-4c2e-a734-74dd69d9195e-kube-api-access-zph6h\") pod \"route-controller-manager-5b5bf74995-db8pv\" (UID: \"85621df9-138d-4c2e-a734-74dd69d9195e\") " pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.351022 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85621df9-138d-4c2e-a734-74dd69d9195e-serving-cert\") pod \"route-controller-manager-5b5bf74995-db8pv\" (UID: \"85621df9-138d-4c2e-a734-74dd69d9195e\") " pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.351112 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85621df9-138d-4c2e-a734-74dd69d9195e-config\") pod \"route-controller-manager-5b5bf74995-db8pv\" (UID: \"85621df9-138d-4c2e-a734-74dd69d9195e\") " pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.351198 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85621df9-138d-4c2e-a734-74dd69d9195e-client-ca\") pod \"route-controller-manager-5b5bf74995-db8pv\" (UID: \"85621df9-138d-4c2e-a734-74dd69d9195e\") " pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.351269 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d9df65-e6cc-404d-8cf9-f50c566d5595-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.351284 4828 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e9d9df65-e6cc-404d-8cf9-f50c566d5595-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.351297 4828 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.351310 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nwbk\" (UniqueName: \"kubernetes.io/projected/e9d9df65-e6cc-404d-8cf9-f50c566d5595-kube-api-access-4nwbk\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.351323 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f498d61-1987-48a1-82d2-14ea6db95c84-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.351334 4828 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.351345 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f498d61-1987-48a1-82d2-14ea6db95c84-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.351355 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9d9df65-e6cc-404d-8cf9-f50c566d5595-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.351365 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fq5cn\" (UniqueName: \"kubernetes.io/projected/5f498d61-1987-48a1-82d2-14ea6db95c84-kube-api-access-fq5cn\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.452837 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zph6h\" (UniqueName: \"kubernetes.io/projected/85621df9-138d-4c2e-a734-74dd69d9195e-kube-api-access-zph6h\") pod \"route-controller-manager-5b5bf74995-db8pv\" (UID: \"85621df9-138d-4c2e-a734-74dd69d9195e\") " pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.453030 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85621df9-138d-4c2e-a734-74dd69d9195e-serving-cert\") pod \"route-controller-manager-5b5bf74995-db8pv\" (UID: \"85621df9-138d-4c2e-a734-74dd69d9195e\") " pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.453126 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85621df9-138d-4c2e-a734-74dd69d9195e-config\") pod \"route-controller-manager-5b5bf74995-db8pv\" (UID: \"85621df9-138d-4c2e-a734-74dd69d9195e\") " pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.453202 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85621df9-138d-4c2e-a734-74dd69d9195e-client-ca\") pod \"route-controller-manager-5b5bf74995-db8pv\" (UID: \"85621df9-138d-4c2e-a734-74dd69d9195e\") " pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.454229 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85621df9-138d-4c2e-a734-74dd69d9195e-client-ca\") pod \"route-controller-manager-5b5bf74995-db8pv\" (UID: \"85621df9-138d-4c2e-a734-74dd69d9195e\") " pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.461358 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85621df9-138d-4c2e-a734-74dd69d9195e-serving-cert\") pod \"route-controller-manager-5b5bf74995-db8pv\" (UID: \"85621df9-138d-4c2e-a734-74dd69d9195e\") " pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.463521 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85621df9-138d-4c2e-a734-74dd69d9195e-config\") pod \"route-controller-manager-5b5bf74995-db8pv\" (UID: \"85621df9-138d-4c2e-a734-74dd69d9195e\") " pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.471125 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zph6h\" (UniqueName: \"kubernetes.io/projected/85621df9-138d-4c2e-a734-74dd69d9195e-kube-api-access-zph6h\") pod \"route-controller-manager-5b5bf74995-db8pv\" (UID: \"85621df9-138d-4c2e-a734-74dd69d9195e\") " pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.512270 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:58:51 crc kubenswrapper[4828]: I1210 18:58:51.677595 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv"] Dec 10 18:58:52 crc kubenswrapper[4828]: I1210 18:58:52.049304 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" event={"ID":"85621df9-138d-4c2e-a734-74dd69d9195e","Type":"ContainerStarted","Data":"ebaed0d648aa0ac1b1615f6d91a92d4633c4659ba4403e376a43563981541f62"} Dec 10 18:58:52 crc kubenswrapper[4828]: I1210 18:58:52.051726 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" event={"ID":"5f498d61-1987-48a1-82d2-14ea6db95c84","Type":"ContainerDied","Data":"f17edbcc305241a3de574eed4317ad4009f9e1ee99d48ee17a481b55a5ad9822"} Dec 10 18:58:52 crc kubenswrapper[4828]: I1210 18:58:52.051779 4828 scope.go:117] "RemoveContainer" containerID="3c38d62467fe683e8ac5876857183ae6676c9a18fa4b9c7c745e83472657f49a" Dec 10 18:58:52 crc kubenswrapper[4828]: I1210 18:58:52.052032 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-n9pnl" Dec 10 18:58:52 crc kubenswrapper[4828]: I1210 18:58:52.054013 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" event={"ID":"e9d9df65-e6cc-404d-8cf9-f50c566d5595","Type":"ContainerDied","Data":"abd3339d25d2b6f80c44f5e8a8eb218c62113ad88f460656a676e9e0fd7c902b"} Dec 10 18:58:52 crc kubenswrapper[4828]: I1210 18:58:52.054151 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2" Dec 10 18:58:52 crc kubenswrapper[4828]: I1210 18:58:52.067911 4828 scope.go:117] "RemoveContainer" containerID="23203906e6c8e9ab5602b14502fb40b91bdc27164c2394aa378d570566a2f1f8" Dec 10 18:58:52 crc kubenswrapper[4828]: I1210 18:58:52.079736 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n9pnl"] Dec 10 18:58:52 crc kubenswrapper[4828]: I1210 18:58:52.085392 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n9pnl"] Dec 10 18:58:52 crc kubenswrapper[4828]: I1210 18:58:52.094100 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2"] Dec 10 18:58:52 crc kubenswrapper[4828]: I1210 18:58:52.098607 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-j8vl2"] Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.251027 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-56f99f9589-4ljp7"] Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.252215 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.255124 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.258483 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.258489 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.260161 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.260217 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.265080 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.266603 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56f99f9589-4ljp7"] Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.267786 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.366419 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" podUID="53029d0b-3ada-4e72-856c-381ab7ae6292" containerName="oauth-openshift" containerID="cri-o://2551383528f4b0b56017f6748a56555e2a0982d52e02d1b15f00d94866a614be" gracePeriod=15 Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.375554 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-client-ca\") pod \"controller-manager-56f99f9589-4ljp7\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.375605 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-config\") pod \"controller-manager-56f99f9589-4ljp7\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.375628 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-proxy-ca-bundles\") pod \"controller-manager-56f99f9589-4ljp7\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.375698 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34219f30-45e9-4be7-9b99-ddc8046fccbe-serving-cert\") pod \"controller-manager-56f99f9589-4ljp7\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.375755 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmx7v\" (UniqueName: \"kubernetes.io/projected/34219f30-45e9-4be7-9b99-ddc8046fccbe-kube-api-access-hmx7v\") pod \"controller-manager-56f99f9589-4ljp7\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.477102 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmx7v\" (UniqueName: \"kubernetes.io/projected/34219f30-45e9-4be7-9b99-ddc8046fccbe-kube-api-access-hmx7v\") pod \"controller-manager-56f99f9589-4ljp7\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.477213 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-client-ca\") pod \"controller-manager-56f99f9589-4ljp7\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.477248 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-config\") pod \"controller-manager-56f99f9589-4ljp7\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.477287 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-proxy-ca-bundles\") pod \"controller-manager-56f99f9589-4ljp7\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.477310 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34219f30-45e9-4be7-9b99-ddc8046fccbe-serving-cert\") pod \"controller-manager-56f99f9589-4ljp7\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.479788 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-client-ca\") pod \"controller-manager-56f99f9589-4ljp7\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.479910 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-config\") pod \"controller-manager-56f99f9589-4ljp7\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.479866 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-proxy-ca-bundles\") pod \"controller-manager-56f99f9589-4ljp7\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.484043 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34219f30-45e9-4be7-9b99-ddc8046fccbe-serving-cert\") pod \"controller-manager-56f99f9589-4ljp7\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.496849 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmx7v\" (UniqueName: \"kubernetes.io/projected/34219f30-45e9-4be7-9b99-ddc8046fccbe-kube-api-access-hmx7v\") pod \"controller-manager-56f99f9589-4ljp7\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.577226 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.795770 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f498d61-1987-48a1-82d2-14ea6db95c84" path="/var/lib/kubelet/pods/5f498d61-1987-48a1-82d2-14ea6db95c84/volumes" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.796655 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9d9df65-e6cc-404d-8cf9-f50c566d5595" path="/var/lib/kubelet/pods/e9d9df65-e6cc-404d-8cf9-f50c566d5595/volumes" Dec 10 18:58:53 crc kubenswrapper[4828]: I1210 18:58:53.973704 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56f99f9589-4ljp7"] Dec 10 18:58:53 crc kubenswrapper[4828]: W1210 18:58:53.977457 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34219f30_45e9_4be7_9b99_ddc8046fccbe.slice/crio-7e96ac58a907e374473c9c38f6d1a4ba130fc2700d32a882c5fe26507525ce42 WatchSource:0}: Error finding container 7e96ac58a907e374473c9c38f6d1a4ba130fc2700d32a882c5fe26507525ce42: Status 404 returned error can't find the container with id 7e96ac58a907e374473c9c38f6d1a4ba130fc2700d32a882c5fe26507525ce42 Dec 10 18:58:54 crc kubenswrapper[4828]: I1210 18:58:54.069631 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" event={"ID":"34219f30-45e9-4be7-9b99-ddc8046fccbe","Type":"ContainerStarted","Data":"7e96ac58a907e374473c9c38f6d1a4ba130fc2700d32a882c5fe26507525ce42"} Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.077302 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5d8xc" event={"ID":"2e51cf27-f229-4c52-b839-c5f957e0fa99","Type":"ContainerStarted","Data":"d92f5cf24f955edfd9ed3b157288034cdce48f76a3a5d17afb4c89f60f80d14d"} Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.079869 4828 generic.go:334] "Generic (PLEG): container finished" podID="53029d0b-3ada-4e72-856c-381ab7ae6292" containerID="2551383528f4b0b56017f6748a56555e2a0982d52e02d1b15f00d94866a614be" exitCode=0 Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.079922 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" event={"ID":"53029d0b-3ada-4e72-856c-381ab7ae6292","Type":"ContainerDied","Data":"2551383528f4b0b56017f6748a56555e2a0982d52e02d1b15f00d94866a614be"} Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.081457 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ll5gw" event={"ID":"44b88622-40eb-4bcf-81a5-dab1f8b4c63b","Type":"ContainerStarted","Data":"a36f4ae4f7d89b2d0700e44d25d272f7e302cc3d88762fe33c8e233f20c584f2"} Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.083310 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2vvgb" event={"ID":"43cd3400-d5e9-4279-9a85-051625ca4607","Type":"ContainerStarted","Data":"566f38d4316412e795518f193b1edf77c2317782cc06cd4dea00d0fa0b5f882f"} Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.085093 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfgtf" event={"ID":"e466f0bb-f597-46e1-b27a-a361322a6dcd","Type":"ContainerStarted","Data":"f38682501ce035cb7c913628c69c4d62174d66e1903f367119c8168eac7d4095"} Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.086710 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" event={"ID":"85621df9-138d-4c2e-a734-74dd69d9195e","Type":"ContainerStarted","Data":"ca4564f70a0189c4cf3f6f88061214e71c5abd095eb76ef0a640ffe69853c08f"} Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.087531 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.088728 4828 patch_prober.go:28] interesting pod/route-controller-manager-5b5bf74995-db8pv container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" start-of-body= Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.088812 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.090165 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs5qt" event={"ID":"b6084fc6-1311-4715-9d6d-01b42fec5f4b","Type":"ContainerStarted","Data":"5779b1692ba90a6ea89eeced9e7d6749c00e508dc24744471fe2efd9ed8fcaf0"} Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.091655 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71" exitCode=0 Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.091740 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71"} Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.094223 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftr4x" event={"ID":"3ef84f18-e802-4038-ba08-2b4eb948d803","Type":"ContainerStarted","Data":"a8862133845c223cf7641c8e31cbade95eddb6d7fda7b09a42edf8bf3e3c86e5"} Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.104907 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" event={"ID":"34219f30-45e9-4be7-9b99-ddc8046fccbe","Type":"ContainerStarted","Data":"5f23016f9587005a00c9501faad940f885a734b0f83a14a9e0b181fa0119cd39"} Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.105407 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.108052 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gs2g5" event={"ID":"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc","Type":"ContainerStarted","Data":"ae1236337f8725a6a27cfcff66f4460525b8ab1833f64f3e8880efd3d63a8aef"} Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.115231 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.118489 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xlcc" event={"ID":"1f95c086-c5d6-4ad2-97e6-79e652fa0040","Type":"ContainerStarted","Data":"f62c9e024dd01b2a6750729b340a9a88f3da3ca61af4f0eb2db395d87ebe232c"} Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.164426 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" podStartSLOduration=14.164408142 podStartE2EDuration="14.164408142s" podCreationTimestamp="2025-12-10 18:58:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:58:55.146209659 +0000 UTC m=+215.656820664" watchObservedRunningTime="2025-12-10 18:58:55.164408142 +0000 UTC m=+215.675019147" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.269918 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" podStartSLOduration=14.26990141 podStartE2EDuration="14.26990141s" podCreationTimestamp="2025-12-10 18:58:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:58:55.243725101 +0000 UTC m=+215.754336126" watchObservedRunningTime="2025-12-10 18:58:55.26990141 +0000 UTC m=+215.780512415" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.827259 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.910695 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-ocp-branding-template\") pod \"53029d0b-3ada-4e72-856c-381ab7ae6292\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.910756 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-error\") pod \"53029d0b-3ada-4e72-856c-381ab7ae6292\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.910792 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9bl4\" (UniqueName: \"kubernetes.io/projected/53029d0b-3ada-4e72-856c-381ab7ae6292-kube-api-access-k9bl4\") pod \"53029d0b-3ada-4e72-856c-381ab7ae6292\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.910839 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-session\") pod \"53029d0b-3ada-4e72-856c-381ab7ae6292\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.910868 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/53029d0b-3ada-4e72-856c-381ab7ae6292-audit-dir\") pod \"53029d0b-3ada-4e72-856c-381ab7ae6292\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.910972 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-serving-cert\") pod \"53029d0b-3ada-4e72-856c-381ab7ae6292\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.911068 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-provider-selection\") pod \"53029d0b-3ada-4e72-856c-381ab7ae6292\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.911097 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-cliconfig\") pod \"53029d0b-3ada-4e72-856c-381ab7ae6292\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.911119 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-router-certs\") pod \"53029d0b-3ada-4e72-856c-381ab7ae6292\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.911143 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-idp-0-file-data\") pod \"53029d0b-3ada-4e72-856c-381ab7ae6292\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.911180 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-service-ca\") pod \"53029d0b-3ada-4e72-856c-381ab7ae6292\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.911201 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-login\") pod \"53029d0b-3ada-4e72-856c-381ab7ae6292\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.911223 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-trusted-ca-bundle\") pod \"53029d0b-3ada-4e72-856c-381ab7ae6292\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.911271 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-audit-policies\") pod \"53029d0b-3ada-4e72-856c-381ab7ae6292\" (UID: \"53029d0b-3ada-4e72-856c-381ab7ae6292\") " Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.911094 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53029d0b-3ada-4e72-856c-381ab7ae6292-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "53029d0b-3ada-4e72-856c-381ab7ae6292" (UID: "53029d0b-3ada-4e72-856c-381ab7ae6292"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.912626 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "53029d0b-3ada-4e72-856c-381ab7ae6292" (UID: "53029d0b-3ada-4e72-856c-381ab7ae6292"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.912664 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "53029d0b-3ada-4e72-856c-381ab7ae6292" (UID: "53029d0b-3ada-4e72-856c-381ab7ae6292"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.913340 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "53029d0b-3ada-4e72-856c-381ab7ae6292" (UID: "53029d0b-3ada-4e72-856c-381ab7ae6292"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.913654 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "53029d0b-3ada-4e72-856c-381ab7ae6292" (UID: "53029d0b-3ada-4e72-856c-381ab7ae6292"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.916949 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "53029d0b-3ada-4e72-856c-381ab7ae6292" (UID: "53029d0b-3ada-4e72-856c-381ab7ae6292"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.917147 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53029d0b-3ada-4e72-856c-381ab7ae6292-kube-api-access-k9bl4" (OuterVolumeSpecName: "kube-api-access-k9bl4") pod "53029d0b-3ada-4e72-856c-381ab7ae6292" (UID: "53029d0b-3ada-4e72-856c-381ab7ae6292"). InnerVolumeSpecName "kube-api-access-k9bl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.917255 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "53029d0b-3ada-4e72-856c-381ab7ae6292" (UID: "53029d0b-3ada-4e72-856c-381ab7ae6292"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.917295 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "53029d0b-3ada-4e72-856c-381ab7ae6292" (UID: "53029d0b-3ada-4e72-856c-381ab7ae6292"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.917736 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "53029d0b-3ada-4e72-856c-381ab7ae6292" (UID: "53029d0b-3ada-4e72-856c-381ab7ae6292"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.917914 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "53029d0b-3ada-4e72-856c-381ab7ae6292" (UID: "53029d0b-3ada-4e72-856c-381ab7ae6292"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.918230 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "53029d0b-3ada-4e72-856c-381ab7ae6292" (UID: "53029d0b-3ada-4e72-856c-381ab7ae6292"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.918911 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "53029d0b-3ada-4e72-856c-381ab7ae6292" (UID: "53029d0b-3ada-4e72-856c-381ab7ae6292"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:58:55 crc kubenswrapper[4828]: I1210 18:58:55.919296 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "53029d0b-3ada-4e72-856c-381ab7ae6292" (UID: "53029d0b-3ada-4e72-856c-381ab7ae6292"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.014743 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.014782 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.014809 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.014823 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.014834 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.014848 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.014860 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.014883 4828 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/53029d0b-3ada-4e72-856c-381ab7ae6292-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.014897 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.014911 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.014922 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9bl4\" (UniqueName: \"kubernetes.io/projected/53029d0b-3ada-4e72-856c-381ab7ae6292-kube-api-access-k9bl4\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.014934 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.014946 4828 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/53029d0b-3ada-4e72-856c-381ab7ae6292-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.014957 4828 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/53029d0b-3ada-4e72-856c-381ab7ae6292-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.126216 4828 generic.go:334] "Generic (PLEG): container finished" podID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" containerID="ae1236337f8725a6a27cfcff66f4460525b8ab1833f64f3e8880efd3d63a8aef" exitCode=0 Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.126285 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gs2g5" event={"ID":"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc","Type":"ContainerDied","Data":"ae1236337f8725a6a27cfcff66f4460525b8ab1833f64f3e8880efd3d63a8aef"} Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.128723 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f95c086-c5d6-4ad2-97e6-79e652fa0040" containerID="f62c9e024dd01b2a6750729b340a9a88f3da3ca61af4f0eb2db395d87ebe232c" exitCode=0 Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.128779 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xlcc" event={"ID":"1f95c086-c5d6-4ad2-97e6-79e652fa0040","Type":"ContainerDied","Data":"f62c9e024dd01b2a6750729b340a9a88f3da3ca61af4f0eb2db395d87ebe232c"} Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.132009 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" event={"ID":"53029d0b-3ada-4e72-856c-381ab7ae6292","Type":"ContainerDied","Data":"f2405e5a507e90e1d2a620454b233a57e51405eeba93c2b31ac3cc8288744f8d"} Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.132066 4828 scope.go:117] "RemoveContainer" containerID="2551383528f4b0b56017f6748a56555e2a0982d52e02d1b15f00d94866a614be" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.132170 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hgsxd" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.136244 4828 generic.go:334] "Generic (PLEG): container finished" podID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" containerID="5779b1692ba90a6ea89eeced9e7d6749c00e508dc24744471fe2efd9ed8fcaf0" exitCode=0 Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.136307 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs5qt" event={"ID":"b6084fc6-1311-4715-9d6d-01b42fec5f4b","Type":"ContainerDied","Data":"5779b1692ba90a6ea89eeced9e7d6749c00e508dc24744471fe2efd9ed8fcaf0"} Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.142107 4828 generic.go:334] "Generic (PLEG): container finished" podID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" containerID="a36f4ae4f7d89b2d0700e44d25d272f7e302cc3d88762fe33c8e233f20c584f2" exitCode=0 Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.142178 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ll5gw" event={"ID":"44b88622-40eb-4bcf-81a5-dab1f8b4c63b","Type":"ContainerDied","Data":"a36f4ae4f7d89b2d0700e44d25d272f7e302cc3d88762fe33c8e233f20c584f2"} Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.145330 4828 generic.go:334] "Generic (PLEG): container finished" podID="3ef84f18-e802-4038-ba08-2b4eb948d803" containerID="a8862133845c223cf7641c8e31cbade95eddb6d7fda7b09a42edf8bf3e3c86e5" exitCode=0 Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.145396 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftr4x" event={"ID":"3ef84f18-e802-4038-ba08-2b4eb948d803","Type":"ContainerDied","Data":"a8862133845c223cf7641c8e31cbade95eddb6d7fda7b09a42edf8bf3e3c86e5"} Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.152150 4828 generic.go:334] "Generic (PLEG): container finished" podID="43cd3400-d5e9-4279-9a85-051625ca4607" containerID="566f38d4316412e795518f193b1edf77c2317782cc06cd4dea00d0fa0b5f882f" exitCode=0 Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.153253 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2vvgb" event={"ID":"43cd3400-d5e9-4279-9a85-051625ca4607","Type":"ContainerDied","Data":"566f38d4316412e795518f193b1edf77c2317782cc06cd4dea00d0fa0b5f882f"} Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.158514 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.236410 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hgsxd"] Dec 10 18:58:56 crc kubenswrapper[4828]: I1210 18:58:56.238735 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hgsxd"] Dec 10 18:58:57 crc kubenswrapper[4828]: I1210 18:58:57.160599 4828 generic.go:334] "Generic (PLEG): container finished" podID="e466f0bb-f597-46e1-b27a-a361322a6dcd" containerID="f38682501ce035cb7c913628c69c4d62174d66e1903f367119c8168eac7d4095" exitCode=0 Dec 10 18:58:57 crc kubenswrapper[4828]: I1210 18:58:57.160679 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfgtf" event={"ID":"e466f0bb-f597-46e1-b27a-a361322a6dcd","Type":"ContainerDied","Data":"f38682501ce035cb7c913628c69c4d62174d66e1903f367119c8168eac7d4095"} Dec 10 18:58:57 crc kubenswrapper[4828]: I1210 18:58:57.164048 4828 generic.go:334] "Generic (PLEG): container finished" podID="2e51cf27-f229-4c52-b839-c5f957e0fa99" containerID="d92f5cf24f955edfd9ed3b157288034cdce48f76a3a5d17afb4c89f60f80d14d" exitCode=0 Dec 10 18:58:57 crc kubenswrapper[4828]: I1210 18:58:57.164122 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5d8xc" event={"ID":"2e51cf27-f229-4c52-b839-c5f957e0fa99","Type":"ContainerDied","Data":"d92f5cf24f955edfd9ed3b157288034cdce48f76a3a5d17afb4c89f60f80d14d"} Dec 10 18:58:57 crc kubenswrapper[4828]: I1210 18:58:57.794659 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53029d0b-3ada-4e72-856c-381ab7ae6292" path="/var/lib/kubelet/pods/53029d0b-3ada-4e72-856c-381ab7ae6292/volumes" Dec 10 18:58:58 crc kubenswrapper[4828]: I1210 18:58:58.189025 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"b8b8c9cc4d09570836d4fe54266d84e397b53b05d934a809d3077cafadbddea8"} Dec 10 18:58:59 crc kubenswrapper[4828]: I1210 18:58:59.196443 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xlcc" event={"ID":"1f95c086-c5d6-4ad2-97e6-79e652fa0040","Type":"ContainerStarted","Data":"43ea3e48374dc6c2aff5c489ad72c1d2572057d86cbf9d439754899b12f03383"} Dec 10 18:58:59 crc kubenswrapper[4828]: I1210 18:58:59.224122 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7xlcc" podStartSLOduration=2.910913508 podStartE2EDuration="1m16.224101319s" podCreationTimestamp="2025-12-10 18:57:43 +0000 UTC" firstStartedPulling="2025-12-10 18:57:45.513019008 +0000 UTC m=+146.023630013" lastFinishedPulling="2025-12-10 18:58:58.826206809 +0000 UTC m=+219.336817824" observedRunningTime="2025-12-10 18:58:59.223093382 +0000 UTC m=+219.733704397" watchObservedRunningTime="2025-12-10 18:58:59.224101319 +0000 UTC m=+219.734712324" Dec 10 18:59:01 crc kubenswrapper[4828]: I1210 18:59:01.412078 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-56f99f9589-4ljp7"] Dec 10 18:59:01 crc kubenswrapper[4828]: I1210 18:59:01.412639 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" containerName="controller-manager" containerID="cri-o://5f23016f9587005a00c9501faad940f885a734b0f83a14a9e0b181fa0119cd39" gracePeriod=30 Dec 10 18:59:01 crc kubenswrapper[4828]: I1210 18:59:01.426560 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv"] Dec 10 18:59:01 crc kubenswrapper[4828]: I1210 18:59:01.426991 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" containerName="route-controller-manager" containerID="cri-o://ca4564f70a0189c4cf3f6f88061214e71c5abd095eb76ef0a640ffe69853c08f" gracePeriod=30 Dec 10 18:59:01 crc kubenswrapper[4828]: I1210 18:59:01.513717 4828 patch_prober.go:28] interesting pod/route-controller-manager-5b5bf74995-db8pv container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" start-of-body= Dec 10 18:59:01 crc kubenswrapper[4828]: I1210 18:59:01.513777 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.221670 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ll5gw" event={"ID":"44b88622-40eb-4bcf-81a5-dab1f8b4c63b","Type":"ContainerStarted","Data":"2a1381f205225da9f5c6e9cc8aa98f6addf271ef62ce9e5dccf0ab9021234baf"} Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.259153 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7d9549f6c-gj4pr"] Dec 10 18:59:03 crc kubenswrapper[4828]: E1210 18:59:03.259501 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53029d0b-3ada-4e72-856c-381ab7ae6292" containerName="oauth-openshift" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.259526 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="53029d0b-3ada-4e72-856c-381ab7ae6292" containerName="oauth-openshift" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.259687 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="53029d0b-3ada-4e72-856c-381ab7ae6292" containerName="oauth-openshift" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.260262 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.263458 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.263524 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.263757 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.263774 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.263969 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.264290 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.264332 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.265637 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.265746 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.265925 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.268122 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.269407 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.278850 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.279502 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.281768 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d9549f6c-gj4pr"] Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.289439 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.412721 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.412787 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.412828 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.412846 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.412887 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.412905 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-user-template-error\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.412919 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.412936 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.412956 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-audit-dir\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.412972 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-session\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.412988 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-audit-policies\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.413008 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-user-template-login\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.413025 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.413048 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxgcr\" (UniqueName: \"kubernetes.io/projected/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-kube-api-access-hxgcr\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.513863 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.513914 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-user-template-error\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.513936 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.513954 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.513972 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-audit-dir\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.513989 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-session\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.514008 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-audit-policies\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.514030 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-user-template-login\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.514051 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.514080 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxgcr\" (UniqueName: \"kubernetes.io/projected/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-kube-api-access-hxgcr\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.514109 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.514133 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.514138 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-audit-dir\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.514152 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.514253 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.515552 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-audit-policies\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.515580 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.516570 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.516642 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.520240 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.520496 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-session\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.521357 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.522040 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.522221 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-user-template-error\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.522066 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.522742 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.524111 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-v4-0-config-user-template-login\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.547904 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxgcr\" (UniqueName: \"kubernetes.io/projected/8ab5a672-80db-49a0-9b83-db5b7a5b4e65-kube-api-access-hxgcr\") pod \"oauth-openshift-7d9549f6c-gj4pr\" (UID: \"8ab5a672-80db-49a0-9b83-db5b7a5b4e65\") " pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.578374 4828 patch_prober.go:28] interesting pod/controller-manager-56f99f9589-4ljp7 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" start-of-body= Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.578459 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" Dec 10 18:59:03 crc kubenswrapper[4828]: I1210 18:59:03.582737 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:04 crc kubenswrapper[4828]: I1210 18:59:04.034394 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7xlcc" Dec 10 18:59:04 crc kubenswrapper[4828]: I1210 18:59:04.035350 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7xlcc" Dec 10 18:59:04 crc kubenswrapper[4828]: I1210 18:59:04.228472 4828 generic.go:334] "Generic (PLEG): container finished" podID="34219f30-45e9-4be7-9b99-ddc8046fccbe" containerID="5f23016f9587005a00c9501faad940f885a734b0f83a14a9e0b181fa0119cd39" exitCode=0 Dec 10 18:59:04 crc kubenswrapper[4828]: I1210 18:59:04.228640 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" event={"ID":"34219f30-45e9-4be7-9b99-ddc8046fccbe","Type":"ContainerDied","Data":"5f23016f9587005a00c9501faad940f885a734b0f83a14a9e0b181fa0119cd39"} Dec 10 18:59:04 crc kubenswrapper[4828]: I1210 18:59:04.230352 4828 generic.go:334] "Generic (PLEG): container finished" podID="85621df9-138d-4c2e-a734-74dd69d9195e" containerID="ca4564f70a0189c4cf3f6f88061214e71c5abd095eb76ef0a640ffe69853c08f" exitCode=0 Dec 10 18:59:04 crc kubenswrapper[4828]: I1210 18:59:04.230393 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" event={"ID":"85621df9-138d-4c2e-a734-74dd69d9195e","Type":"ContainerDied","Data":"ca4564f70a0189c4cf3f6f88061214e71c5abd095eb76ef0a640ffe69853c08f"} Dec 10 18:59:04 crc kubenswrapper[4828]: I1210 18:59:04.256576 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ll5gw" podStartSLOduration=5.43800693 podStartE2EDuration="1m20.256559282s" podCreationTimestamp="2025-12-10 18:57:44 +0000 UTC" firstStartedPulling="2025-12-10 18:57:45.505210111 +0000 UTC m=+146.015821126" lastFinishedPulling="2025-12-10 18:59:00.323762453 +0000 UTC m=+220.834373478" observedRunningTime="2025-12-10 18:59:04.255494003 +0000 UTC m=+224.766105008" watchObservedRunningTime="2025-12-10 18:59:04.256559282 +0000 UTC m=+224.767170287" Dec 10 18:59:04 crc kubenswrapper[4828]: I1210 18:59:04.389023 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ll5gw" Dec 10 18:59:04 crc kubenswrapper[4828]: I1210 18:59:04.389088 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ll5gw" Dec 10 18:59:04 crc kubenswrapper[4828]: I1210 18:59:04.426535 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7xlcc" Dec 10 18:59:05 crc kubenswrapper[4828]: I1210 18:59:05.307405 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7xlcc" Dec 10 18:59:05 crc kubenswrapper[4828]: I1210 18:59:05.447536 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-ll5gw" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" containerName="registry-server" probeResult="failure" output=< Dec 10 18:59:05 crc kubenswrapper[4828]: timeout: failed to connect service ":50051" within 1s Dec 10 18:59:05 crc kubenswrapper[4828]: > Dec 10 18:59:11 crc kubenswrapper[4828]: E1210 18:59:11.074722 4828 file.go:109] "Unable to process watch event" err="can't process config file \"/etc/kubernetes/manifests/kube-apiserver-startup-monitor-pod.yaml\": /etc/kubernetes/manifests/kube-apiserver-startup-monitor-pod.yaml: couldn't parse as pod(Object 'Kind' is missing in 'null'), please check config file" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.076172 4828 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.077172 4828 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.077313 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.077593 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2" gracePeriod=15 Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.077708 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9" gracePeriod=15 Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.077731 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f" gracePeriod=15 Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.077752 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50" gracePeriod=15 Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.077706 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203" gracePeriod=15 Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.078585 4828 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 18:59:11 crc kubenswrapper[4828]: E1210 18:59:11.078712 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.078722 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 10 18:59:11 crc kubenswrapper[4828]: E1210 18:59:11.078733 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.078740 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 10 18:59:11 crc kubenswrapper[4828]: E1210 18:59:11.078750 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.078757 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 10 18:59:11 crc kubenswrapper[4828]: E1210 18:59:11.078763 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.078777 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 18:59:11 crc kubenswrapper[4828]: E1210 18:59:11.078786 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.078792 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 10 18:59:11 crc kubenswrapper[4828]: E1210 18:59:11.078812 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.078818 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 10 18:59:11 crc kubenswrapper[4828]: E1210 18:59:11.078824 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.078833 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.078938 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.078951 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.078962 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.078970 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.078979 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.078989 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.216407 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.216476 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.216539 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.216589 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.216634 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.216676 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.216694 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.216723 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.318196 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.318264 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.318308 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.318306 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.318341 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.318381 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.318413 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.318409 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.318457 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.318470 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.318486 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.318513 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.318521 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.318556 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.318593 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.318633 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:11 crc kubenswrapper[4828]: E1210 18:59:11.942046 4828 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:11 crc kubenswrapper[4828]: E1210 18:59:11.943067 4828 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:11 crc kubenswrapper[4828]: E1210 18:59:11.943936 4828 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:11 crc kubenswrapper[4828]: E1210 18:59:11.944735 4828 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:11 crc kubenswrapper[4828]: E1210 18:59:11.945575 4828 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:11 crc kubenswrapper[4828]: I1210 18:59:11.945770 4828 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 10 18:59:11 crc kubenswrapper[4828]: E1210 18:59:11.946307 4828 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="200ms" Dec 10 18:59:12 crc kubenswrapper[4828]: E1210 18:59:12.147403 4828 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="400ms" Dec 10 18:59:12 crc kubenswrapper[4828]: I1210 18:59:12.291647 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 10 18:59:12 crc kubenswrapper[4828]: I1210 18:59:12.293244 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 18:59:12 crc kubenswrapper[4828]: I1210 18:59:12.294156 4828 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50" exitCode=2 Dec 10 18:59:12 crc kubenswrapper[4828]: I1210 18:59:12.512650 4828 patch_prober.go:28] interesting pod/route-controller-manager-5b5bf74995-db8pv container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.56:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 10 18:59:12 crc kubenswrapper[4828]: I1210 18:59:12.512727 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.56:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 10 18:59:12 crc kubenswrapper[4828]: E1210 18:59:12.513281 4828 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/events\": dial tcp 38.102.83.214:6443: connect: connection refused" event=< Dec 10 18:59:12 crc kubenswrapper[4828]: &Event{ObjectMeta:{route-controller-manager-5b5bf74995-db8pv.187fefb8006d5696 openshift-route-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-route-controller-manager,Name:route-controller-manager-5b5bf74995-db8pv,UID:85621df9-138d-4c2e-a734-74dd69d9195e,APIVersion:v1,ResourceVersion:29209,FieldPath:spec.containers{route-controller-manager},},Reason:ProbeError,Message:Readiness probe error: Get "https://10.217.0.56:8443/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Dec 10 18:59:12 crc kubenswrapper[4828]: body: Dec 10 18:59:12 crc kubenswrapper[4828]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 18:59:12.512710294 +0000 UTC m=+233.023321299,LastTimestamp:2025-12-10 18:59:12.512710294 +0000 UTC m=+233.023321299,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 10 18:59:12 crc kubenswrapper[4828]: > Dec 10 18:59:12 crc kubenswrapper[4828]: E1210 18:59:12.548863 4828 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="800ms" Dec 10 18:59:13 crc kubenswrapper[4828]: E1210 18:59:13.350548 4828 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="1.6s" Dec 10 18:59:14 crc kubenswrapper[4828]: I1210 18:59:14.428721 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ll5gw" Dec 10 18:59:14 crc kubenswrapper[4828]: I1210 18:59:14.429207 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:14 crc kubenswrapper[4828]: I1210 18:59:14.464127 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ll5gw" Dec 10 18:59:14 crc kubenswrapper[4828]: I1210 18:59:14.464526 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:14 crc kubenswrapper[4828]: I1210 18:59:14.578000 4828 patch_prober.go:28] interesting pod/controller-manager-56f99f9589-4ljp7 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 10 18:59:14 crc kubenswrapper[4828]: I1210 18:59:14.578073 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 10 18:59:14 crc kubenswrapper[4828]: E1210 18:59:14.951910 4828 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="3.2s" Dec 10 18:59:15 crc kubenswrapper[4828]: I1210 18:59:15.312399 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 10 18:59:15 crc kubenswrapper[4828]: I1210 18:59:15.314213 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 18:59:15 crc kubenswrapper[4828]: I1210 18:59:15.315349 4828 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f" exitCode=0 Dec 10 18:59:15 crc kubenswrapper[4828]: I1210 18:59:15.315446 4828 scope.go:117] "RemoveContainer" containerID="bb9675f4e9a1d07cb66ccecdc13ba7ec4d718cc2df7251e204e6e17273fbc859" Dec 10 18:59:16 crc kubenswrapper[4828]: E1210 18:59:16.104832 4828 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.214:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:16 crc kubenswrapper[4828]: I1210 18:59:16.105507 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:17 crc kubenswrapper[4828]: E1210 18:59:17.021056 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:59:17Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:59:17Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:59:17Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-10T18:59:17Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:17 crc kubenswrapper[4828]: E1210 18:59:17.021346 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:17 crc kubenswrapper[4828]: E1210 18:59:17.021537 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:17 crc kubenswrapper[4828]: E1210 18:59:17.021699 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:17 crc kubenswrapper[4828]: E1210 18:59:17.021963 4828 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:17 crc kubenswrapper[4828]: E1210 18:59:17.021994 4828 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 10 18:59:17 crc kubenswrapper[4828]: I1210 18:59:17.329575 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 18:59:17 crc kubenswrapper[4828]: I1210 18:59:17.330477 4828 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203" exitCode=0 Dec 10 18:59:17 crc kubenswrapper[4828]: I1210 18:59:17.330504 4828 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9" exitCode=0 Dec 10 18:59:17 crc kubenswrapper[4828]: I1210 18:59:17.330511 4828 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2" exitCode=0 Dec 10 18:59:17 crc kubenswrapper[4828]: I1210 18:59:17.332650 4828 generic.go:334] "Generic (PLEG): container finished" podID="867406bf-a6e6-4644-a18b-4fa778cea1a5" containerID="7e0bf15da9681106df3d2f8e11a52c960ea641077bfc3c112b437708170c197e" exitCode=0 Dec 10 18:59:17 crc kubenswrapper[4828]: I1210 18:59:17.332706 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"867406bf-a6e6-4644-a18b-4fa778cea1a5","Type":"ContainerDied","Data":"7e0bf15da9681106df3d2f8e11a52c960ea641077bfc3c112b437708170c197e"} Dec 10 18:59:17 crc kubenswrapper[4828]: I1210 18:59:17.333342 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:17 crc kubenswrapper[4828]: I1210 18:59:17.333611 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:18 crc kubenswrapper[4828]: E1210 18:59:18.153850 4828 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="6.4s" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.714207 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.715091 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.715483 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.715768 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.718536 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.719002 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.719279 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.719522 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.719833 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.797456 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.798101 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.798623 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.799081 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.832928 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zph6h\" (UniqueName: \"kubernetes.io/projected/85621df9-138d-4c2e-a734-74dd69d9195e-kube-api-access-zph6h\") pod \"85621df9-138d-4c2e-a734-74dd69d9195e\" (UID: \"85621df9-138d-4c2e-a734-74dd69d9195e\") " Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.832991 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85621df9-138d-4c2e-a734-74dd69d9195e-config\") pod \"85621df9-138d-4c2e-a734-74dd69d9195e\" (UID: \"85621df9-138d-4c2e-a734-74dd69d9195e\") " Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.833039 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85621df9-138d-4c2e-a734-74dd69d9195e-serving-cert\") pod \"85621df9-138d-4c2e-a734-74dd69d9195e\" (UID: \"85621df9-138d-4c2e-a734-74dd69d9195e\") " Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.833120 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmx7v\" (UniqueName: \"kubernetes.io/projected/34219f30-45e9-4be7-9b99-ddc8046fccbe-kube-api-access-hmx7v\") pod \"34219f30-45e9-4be7-9b99-ddc8046fccbe\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.833208 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85621df9-138d-4c2e-a734-74dd69d9195e-client-ca\") pod \"85621df9-138d-4c2e-a734-74dd69d9195e\" (UID: \"85621df9-138d-4c2e-a734-74dd69d9195e\") " Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.833239 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-config\") pod \"34219f30-45e9-4be7-9b99-ddc8046fccbe\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.833273 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34219f30-45e9-4be7-9b99-ddc8046fccbe-serving-cert\") pod \"34219f30-45e9-4be7-9b99-ddc8046fccbe\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.833305 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-proxy-ca-bundles\") pod \"34219f30-45e9-4be7-9b99-ddc8046fccbe\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.833339 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-client-ca\") pod \"34219f30-45e9-4be7-9b99-ddc8046fccbe\" (UID: \"34219f30-45e9-4be7-9b99-ddc8046fccbe\") " Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.834398 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-client-ca" (OuterVolumeSpecName: "client-ca") pod "34219f30-45e9-4be7-9b99-ddc8046fccbe" (UID: "34219f30-45e9-4be7-9b99-ddc8046fccbe"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.834381 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85621df9-138d-4c2e-a734-74dd69d9195e-config" (OuterVolumeSpecName: "config") pod "85621df9-138d-4c2e-a734-74dd69d9195e" (UID: "85621df9-138d-4c2e-a734-74dd69d9195e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.834534 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85621df9-138d-4c2e-a734-74dd69d9195e-client-ca" (OuterVolumeSpecName: "client-ca") pod "85621df9-138d-4c2e-a734-74dd69d9195e" (UID: "85621df9-138d-4c2e-a734-74dd69d9195e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.834597 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "34219f30-45e9-4be7-9b99-ddc8046fccbe" (UID: "34219f30-45e9-4be7-9b99-ddc8046fccbe"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.834699 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-config" (OuterVolumeSpecName: "config") pod "34219f30-45e9-4be7-9b99-ddc8046fccbe" (UID: "34219f30-45e9-4be7-9b99-ddc8046fccbe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.840207 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34219f30-45e9-4be7-9b99-ddc8046fccbe-kube-api-access-hmx7v" (OuterVolumeSpecName: "kube-api-access-hmx7v") pod "34219f30-45e9-4be7-9b99-ddc8046fccbe" (UID: "34219f30-45e9-4be7-9b99-ddc8046fccbe"). InnerVolumeSpecName "kube-api-access-hmx7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.840620 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85621df9-138d-4c2e-a734-74dd69d9195e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "85621df9-138d-4c2e-a734-74dd69d9195e" (UID: "85621df9-138d-4c2e-a734-74dd69d9195e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.840625 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85621df9-138d-4c2e-a734-74dd69d9195e-kube-api-access-zph6h" (OuterVolumeSpecName: "kube-api-access-zph6h") pod "85621df9-138d-4c2e-a734-74dd69d9195e" (UID: "85621df9-138d-4c2e-a734-74dd69d9195e"). InnerVolumeSpecName "kube-api-access-zph6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.841341 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34219f30-45e9-4be7-9b99-ddc8046fccbe-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "34219f30-45e9-4be7-9b99-ddc8046fccbe" (UID: "34219f30-45e9-4be7-9b99-ddc8046fccbe"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.934424 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmx7v\" (UniqueName: \"kubernetes.io/projected/34219f30-45e9-4be7-9b99-ddc8046fccbe-kube-api-access-hmx7v\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.934462 4828 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85621df9-138d-4c2e-a734-74dd69d9195e-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.934473 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.934482 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34219f30-45e9-4be7-9b99-ddc8046fccbe-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.934491 4828 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.934500 4828 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/34219f30-45e9-4be7-9b99-ddc8046fccbe-client-ca\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.934511 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zph6h\" (UniqueName: \"kubernetes.io/projected/85621df9-138d-4c2e-a734-74dd69d9195e-kube-api-access-zph6h\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.934522 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85621df9-138d-4c2e-a734-74dd69d9195e-config\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:19 crc kubenswrapper[4828]: I1210 18:59:19.934532 4828 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85621df9-138d-4c2e-a734-74dd69d9195e-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.353025 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" event={"ID":"85621df9-138d-4c2e-a734-74dd69d9195e","Type":"ContainerDied","Data":"ebaed0d648aa0ac1b1615f6d91a92d4633c4659ba4403e376a43563981541f62"} Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.353151 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.354010 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.354376 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.354911 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.354995 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" event={"ID":"34219f30-45e9-4be7-9b99-ddc8046fccbe","Type":"ContainerDied","Data":"7e96ac58a907e374473c9c38f6d1a4ba130fc2700d32a882c5fe26507525ce42"} Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.355081 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.355549 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.356165 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.356580 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.357161 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.357565 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.371089 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.371494 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.371931 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.372319 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.380709 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.381466 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.382001 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:20 crc kubenswrapper[4828]: I1210 18:59:20.382416 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:21 crc kubenswrapper[4828]: E1210 18:59:21.234265 4828 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/events\": dial tcp 38.102.83.214:6443: connect: connection refused" event=< Dec 10 18:59:21 crc kubenswrapper[4828]: &Event{ObjectMeta:{route-controller-manager-5b5bf74995-db8pv.187fefb8006d5696 openshift-route-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-route-controller-manager,Name:route-controller-manager-5b5bf74995-db8pv,UID:85621df9-138d-4c2e-a734-74dd69d9195e,APIVersion:v1,ResourceVersion:29209,FieldPath:spec.containers{route-controller-manager},},Reason:ProbeError,Message:Readiness probe error: Get "https://10.217.0.56:8443/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Dec 10 18:59:21 crc kubenswrapper[4828]: body: Dec 10 18:59:21 crc kubenswrapper[4828]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-10 18:59:12.512710294 +0000 UTC m=+233.023321299,LastTimestamp:2025-12-10 18:59:12.512710294 +0000 UTC m=+233.023321299,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 10 18:59:21 crc kubenswrapper[4828]: > Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.601531 4828 scope.go:117] "RemoveContainer" containerID="ca4564f70a0189c4cf3f6f88061214e71c5abd095eb76ef0a640ffe69853c08f" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.620147 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.620932 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.621528 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.621772 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.621994 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.628184 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.629396 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.630003 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.630409 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.631209 4828 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.631404 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.631588 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:23 crc kubenswrapper[4828]: W1210 18:59:23.660847 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-def7bc69e74757d5f71a31b5351910286f1121773995bb23241de3a760e742bd WatchSource:0}: Error finding container def7bc69e74757d5f71a31b5351910286f1121773995bb23241de3a760e742bd: Status 404 returned error can't find the container with id def7bc69e74757d5f71a31b5351910286f1121773995bb23241de3a760e742bd Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.686418 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/867406bf-a6e6-4644-a18b-4fa778cea1a5-kube-api-access\") pod \"867406bf-a6e6-4644-a18b-4fa778cea1a5\" (UID: \"867406bf-a6e6-4644-a18b-4fa778cea1a5\") " Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.686456 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.686472 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.686496 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.686530 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/867406bf-a6e6-4644-a18b-4fa778cea1a5-var-lock\") pod \"867406bf-a6e6-4644-a18b-4fa778cea1a5\" (UID: \"867406bf-a6e6-4644-a18b-4fa778cea1a5\") " Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.686626 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/867406bf-a6e6-4644-a18b-4fa778cea1a5-kubelet-dir\") pod \"867406bf-a6e6-4644-a18b-4fa778cea1a5\" (UID: \"867406bf-a6e6-4644-a18b-4fa778cea1a5\") " Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.686883 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/867406bf-a6e6-4644-a18b-4fa778cea1a5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "867406bf-a6e6-4644-a18b-4fa778cea1a5" (UID: "867406bf-a6e6-4644-a18b-4fa778cea1a5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.686878 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.686934 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.686956 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/867406bf-a6e6-4644-a18b-4fa778cea1a5-var-lock" (OuterVolumeSpecName: "var-lock") pod "867406bf-a6e6-4644-a18b-4fa778cea1a5" (UID: "867406bf-a6e6-4644-a18b-4fa778cea1a5"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.686980 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.687666 4828 scope.go:117] "RemoveContainer" containerID="5f23016f9587005a00c9501faad940f885a734b0f83a14a9e0b181fa0119cd39" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.689330 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/867406bf-a6e6-4644-a18b-4fa778cea1a5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "867406bf-a6e6-4644-a18b-4fa778cea1a5" (UID: "867406bf-a6e6-4644-a18b-4fa778cea1a5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.789922 4828 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.790159 4828 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.790170 4828 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/867406bf-a6e6-4644-a18b-4fa778cea1a5-var-lock\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.790179 4828 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/867406bf-a6e6-4644-a18b-4fa778cea1a5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.790187 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/867406bf-a6e6-4644-a18b-4fa778cea1a5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.790197 4828 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 10 18:59:23 crc kubenswrapper[4828]: I1210 18:59:23.806447 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 10 18:59:24 crc kubenswrapper[4828]: E1210 18:59:24.209251 4828 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 10 18:59:24 crc kubenswrapper[4828]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-7d9549f6c-gj4pr_openshift-authentication_8ab5a672-80db-49a0-9b83-db5b7a5b4e65_0(c340714623d3d61eb6a0614fc240a1a2874b3c64a708539db8287953d1557e54): error adding pod openshift-authentication_oauth-openshift-7d9549f6c-gj4pr to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"c340714623d3d61eb6a0614fc240a1a2874b3c64a708539db8287953d1557e54" Netns:"/var/run/netns/59856d45-910e-4829-8d70-0920509722c4" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-7d9549f6c-gj4pr;K8S_POD_INFRA_CONTAINER_ID=c340714623d3d61eb6a0614fc240a1a2874b3c64a708539db8287953d1557e54;K8S_POD_UID=8ab5a672-80db-49a0-9b83-db5b7a5b4e65" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-7d9549f6c-gj4pr] networking: Multus: [openshift-authentication/oauth-openshift-7d9549f6c-gj4pr/8ab5a672-80db-49a0-9b83-db5b7a5b4e65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod oauth-openshift-7d9549f6c-gj4pr in out of cluster comm: SetNetworkStatus: failed to update the pod oauth-openshift-7d9549f6c-gj4pr in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7d9549f6c-gj4pr?timeout=1m0s": dial tcp 38.102.83.214:6443: connect: connection refused Dec 10 18:59:24 crc kubenswrapper[4828]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 10 18:59:24 crc kubenswrapper[4828]: > Dec 10 18:59:24 crc kubenswrapper[4828]: E1210 18:59:24.209648 4828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 10 18:59:24 crc kubenswrapper[4828]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-7d9549f6c-gj4pr_openshift-authentication_8ab5a672-80db-49a0-9b83-db5b7a5b4e65_0(c340714623d3d61eb6a0614fc240a1a2874b3c64a708539db8287953d1557e54): error adding pod openshift-authentication_oauth-openshift-7d9549f6c-gj4pr to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"c340714623d3d61eb6a0614fc240a1a2874b3c64a708539db8287953d1557e54" Netns:"/var/run/netns/59856d45-910e-4829-8d70-0920509722c4" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-7d9549f6c-gj4pr;K8S_POD_INFRA_CONTAINER_ID=c340714623d3d61eb6a0614fc240a1a2874b3c64a708539db8287953d1557e54;K8S_POD_UID=8ab5a672-80db-49a0-9b83-db5b7a5b4e65" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-7d9549f6c-gj4pr] networking: Multus: [openshift-authentication/oauth-openshift-7d9549f6c-gj4pr/8ab5a672-80db-49a0-9b83-db5b7a5b4e65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod oauth-openshift-7d9549f6c-gj4pr in out of cluster comm: SetNetworkStatus: failed to update the pod oauth-openshift-7d9549f6c-gj4pr in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7d9549f6c-gj4pr?timeout=1m0s": dial tcp 38.102.83.214:6443: connect: connection refused Dec 10 18:59:24 crc kubenswrapper[4828]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 10 18:59:24 crc kubenswrapper[4828]: > pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:24 crc kubenswrapper[4828]: E1210 18:59:24.209680 4828 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 10 18:59:24 crc kubenswrapper[4828]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-7d9549f6c-gj4pr_openshift-authentication_8ab5a672-80db-49a0-9b83-db5b7a5b4e65_0(c340714623d3d61eb6a0614fc240a1a2874b3c64a708539db8287953d1557e54): error adding pod openshift-authentication_oauth-openshift-7d9549f6c-gj4pr to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"c340714623d3d61eb6a0614fc240a1a2874b3c64a708539db8287953d1557e54" Netns:"/var/run/netns/59856d45-910e-4829-8d70-0920509722c4" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-7d9549f6c-gj4pr;K8S_POD_INFRA_CONTAINER_ID=c340714623d3d61eb6a0614fc240a1a2874b3c64a708539db8287953d1557e54;K8S_POD_UID=8ab5a672-80db-49a0-9b83-db5b7a5b4e65" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-7d9549f6c-gj4pr] networking: Multus: [openshift-authentication/oauth-openshift-7d9549f6c-gj4pr/8ab5a672-80db-49a0-9b83-db5b7a5b4e65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod oauth-openshift-7d9549f6c-gj4pr in out of cluster comm: SetNetworkStatus: failed to update the pod oauth-openshift-7d9549f6c-gj4pr in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7d9549f6c-gj4pr?timeout=1m0s": dial tcp 38.102.83.214:6443: connect: connection refused Dec 10 18:59:24 crc kubenswrapper[4828]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 10 18:59:24 crc kubenswrapper[4828]: > pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:24 crc kubenswrapper[4828]: E1210 18:59:24.209752 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"oauth-openshift-7d9549f6c-gj4pr_openshift-authentication(8ab5a672-80db-49a0-9b83-db5b7a5b4e65)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"oauth-openshift-7d9549f6c-gj4pr_openshift-authentication(8ab5a672-80db-49a0-9b83-db5b7a5b4e65)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-7d9549f6c-gj4pr_openshift-authentication_8ab5a672-80db-49a0-9b83-db5b7a5b4e65_0(c340714623d3d61eb6a0614fc240a1a2874b3c64a708539db8287953d1557e54): error adding pod openshift-authentication_oauth-openshift-7d9549f6c-gj4pr to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"c340714623d3d61eb6a0614fc240a1a2874b3c64a708539db8287953d1557e54\\\" Netns:\\\"/var/run/netns/59856d45-910e-4829-8d70-0920509722c4\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-7d9549f6c-gj4pr;K8S_POD_INFRA_CONTAINER_ID=c340714623d3d61eb6a0614fc240a1a2874b3c64a708539db8287953d1557e54;K8S_POD_UID=8ab5a672-80db-49a0-9b83-db5b7a5b4e65\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-7d9549f6c-gj4pr] networking: Multus: [openshift-authentication/oauth-openshift-7d9549f6c-gj4pr/8ab5a672-80db-49a0-9b83-db5b7a5b4e65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod oauth-openshift-7d9549f6c-gj4pr in out of cluster comm: SetNetworkStatus: failed to update the pod oauth-openshift-7d9549f6c-gj4pr in out of cluster comm: status update failed for pod /: Get \\\"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7d9549f6c-gj4pr?timeout=1m0s\\\": dial tcp 38.102.83.214:6443: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" podUID="8ab5a672-80db-49a0-9b83-db5b7a5b4e65" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.382052 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs5qt" event={"ID":"b6084fc6-1311-4715-9d6d-01b42fec5f4b","Type":"ContainerStarted","Data":"b32875c68b7888406fadd752847829bdc52c11da3f660b23e7a1d0d41c6f2c3b"} Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.382766 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.383166 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.383373 4828 status_manager.go:851] "Failed to get status for pod" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" pod="openshift-marketplace/redhat-marketplace-qs5qt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-qs5qt\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.383607 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.383854 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.384916 4828 scope.go:117] "RemoveContainer" containerID="cc16cada1cd8a7ad9e339ee07f8fab53eccd997340d956c9d06812aaecc0265f" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.384930 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.385349 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.385736 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.385991 4828 status_manager.go:851] "Failed to get status for pod" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" pod="openshift-marketplace/redhat-marketplace-qs5qt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-qs5qt\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.386178 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.386746 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.386903 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"867406bf-a6e6-4644-a18b-4fa778cea1a5","Type":"ContainerDied","Data":"960167c72cf91fc50de7ec6b94de36dead12bbc4fe9ec62e1e0facb867ddd855"} Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.386928 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="960167c72cf91fc50de7ec6b94de36dead12bbc4fe9ec62e1e0facb867ddd855" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.386965 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.387013 4828 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.388225 4828 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.388476 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.388776 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.389065 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.389378 4828 status_manager.go:851] "Failed to get status for pod" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" pod="openshift-marketplace/redhat-marketplace-qs5qt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-qs5qt\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.389586 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.390097 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.390129 4828 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6" exitCode=1 Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.390160 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.390179 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6"} Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.390355 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.390513 4828 status_manager.go:851] "Failed to get status for pod" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" pod="openshift-marketplace/redhat-marketplace-qs5qt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-qs5qt\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.390586 4828 scope.go:117] "RemoveContainer" containerID="54410aa38ce08e4ac855ec6c8c3e217121f31af175c4c77adcf64ac3689cbef6" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.390658 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.390823 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.391144 4828 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.391370 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.391523 4828 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.391663 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.392293 4828 status_manager.go:851] "Failed to get status for pod" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" pod="openshift-marketplace/redhat-marketplace-qs5qt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-qs5qt\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.392668 4828 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.393012 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.393192 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.394331 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"1c65abebabfa38770059d8a3ee3bad6c5cafea58415c02a7eaf3efcbbdbb797c"} Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.394370 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"def7bc69e74757d5f71a31b5351910286f1121773995bb23241de3a760e742bd"} Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.394694 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.394870 4828 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: E1210 18:59:24.394922 4828 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.214:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.395020 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.395152 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.395296 4828 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.396327 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.396987 4828 status_manager.go:851] "Failed to get status for pod" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" pod="openshift-marketplace/redhat-marketplace-qs5qt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-qs5qt\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.400667 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gs2g5" event={"ID":"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc","Type":"ContainerStarted","Data":"42e14c173fa2164db28101ad65272161516c6ce98e5793101e69699ba00e4747"} Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.402433 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.402941 4828 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.403225 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.403458 4828 status_manager.go:851] "Failed to get status for pod" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" pod="openshift-marketplace/redhat-marketplace-qs5qt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-qs5qt\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.403753 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.403973 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.404118 4828 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.404259 4828 status_manager.go:851] "Failed to get status for pod" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" pod="openshift-marketplace/certified-operators-gs2g5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gs2g5\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.406687 4828 scope.go:117] "RemoveContainer" containerID="95a309565554bbf073cdf99c5716458905051abdd76243818faffff715e21203" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.409242 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfgtf" event={"ID":"e466f0bb-f597-46e1-b27a-a361322a6dcd","Type":"ContainerStarted","Data":"e5acfe41e8f64a9ae2b49b3c66e8f1b9b3940b390f2a934974073d6e4afee137"} Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.410265 4828 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.410725 4828 status_manager.go:851] "Failed to get status for pod" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" pod="openshift-marketplace/certified-operators-gs2g5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gs2g5\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.411979 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.412326 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.412495 4828 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.412673 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.412851 4828 status_manager.go:851] "Failed to get status for pod" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" pod="openshift-marketplace/redhat-marketplace-qs5qt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-qs5qt\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.413199 4828 status_manager.go:851] "Failed to get status for pod" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" pod="openshift-marketplace/redhat-operators-kfgtf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kfgtf\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.413669 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.416961 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2vvgb" event={"ID":"43cd3400-d5e9-4279-9a85-051625ca4607","Type":"ContainerStarted","Data":"b64b11b0c4a38dfdbb4d958370a441a18ed178d13810ae372fca0e3827a6cc80"} Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.418350 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.418908 4828 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.419116 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.419418 4828 status_manager.go:851] "Failed to get status for pod" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" pod="openshift-marketplace/redhat-marketplace-qs5qt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-qs5qt\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.419890 4828 status_manager.go:851] "Failed to get status for pod" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" pod="openshift-marketplace/redhat-operators-kfgtf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kfgtf\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.420079 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.420332 4828 status_manager.go:851] "Failed to get status for pod" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" pod="openshift-marketplace/certified-operators-2vvgb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2vvgb\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.420677 4828 status_manager.go:851] "Failed to get status for pod" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" pod="openshift-marketplace/certified-operators-gs2g5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gs2g5\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.420898 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.421068 4828 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.422321 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5d8xc" event={"ID":"2e51cf27-f229-4c52-b839-c5f957e0fa99","Type":"ContainerStarted","Data":"db2abf9b0f92038dc8c11083bf5dabea39754b8c6e01f571bf1429744d72c324"} Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.423310 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.423581 4828 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.423947 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.424151 4828 status_manager.go:851] "Failed to get status for pod" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" pod="openshift-marketplace/redhat-marketplace-qs5qt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-qs5qt\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.424402 4828 status_manager.go:851] "Failed to get status for pod" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" pod="openshift-marketplace/redhat-operators-5d8xc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5d8xc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.424706 4828 status_manager.go:851] "Failed to get status for pod" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" pod="openshift-marketplace/redhat-operators-kfgtf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kfgtf\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.424897 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.425051 4828 status_manager.go:851] "Failed to get status for pod" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" pod="openshift-marketplace/certified-operators-2vvgb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2vvgb\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.425198 4828 status_manager.go:851] "Failed to get status for pod" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" pod="openshift-marketplace/certified-operators-gs2g5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gs2g5\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.425439 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.425825 4828 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.436396 4828 scope.go:117] "RemoveContainer" containerID="d8eff21b9a1155f92da6c98cac0f93823e9a3be34927d5a72213b3e615a552a9" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.437348 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftr4x" event={"ID":"3ef84f18-e802-4038-ba08-2b4eb948d803","Type":"ContainerStarted","Data":"40cba23895e88153418f64c22b79e8dd857b8c34c57d2e037eedd2053ebe9b65"} Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.438435 4828 status_manager.go:851] "Failed to get status for pod" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" pod="openshift-marketplace/certified-operators-2vvgb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2vvgb\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.438602 4828 status_manager.go:851] "Failed to get status for pod" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" pod="openshift-marketplace/certified-operators-gs2g5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gs2g5\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.438747 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.439025 4828 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.439618 4828 status_manager.go:851] "Failed to get status for pod" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" pod="openshift-marketplace/redhat-marketplace-ftr4x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ftr4x\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.440177 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.440367 4828 status_manager.go:851] "Failed to get status for pod" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" pod="openshift-marketplace/redhat-marketplace-qs5qt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-qs5qt\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.440517 4828 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.440529 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.440675 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.440876 4828 status_manager.go:851] "Failed to get status for pod" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" pod="openshift-marketplace/redhat-operators-kfgtf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kfgtf\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.441048 4828 status_manager.go:851] "Failed to get status for pod" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" pod="openshift-marketplace/redhat-operators-5d8xc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5d8xc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.441223 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.441393 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.456847 4828 scope.go:117] "RemoveContainer" containerID="c5c362c445d0be468cce66b7dff983c0c7f7b1086e6f75ba2125001186f9fc50" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.471144 4828 scope.go:117] "RemoveContainer" containerID="5b164e2a9cc97d9b53367f15619840a391df89017dd20f93e10099be3235ead2" Dec 10 18:59:24 crc kubenswrapper[4828]: I1210 18:59:24.506563 4828 scope.go:117] "RemoveContainer" containerID="364677057c6d0d033a92ba483430593288a8b6eb78c9f283d967c9ea4be1e530" Dec 10 18:59:24 crc kubenswrapper[4828]: E1210 18:59:24.555231 4828 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.214:6443: connect: connection refused" interval="7s" Dec 10 18:59:24 crc kubenswrapper[4828]: E1210 18:59:24.880640 4828 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 10 18:59:24 crc kubenswrapper[4828]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-7d9549f6c-gj4pr_openshift-authentication_8ab5a672-80db-49a0-9b83-db5b7a5b4e65_0(1fc32ef2e5384487a770b223961b9ed6147aa70f30b1718460e76d17160c1183): error adding pod openshift-authentication_oauth-openshift-7d9549f6c-gj4pr to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"1fc32ef2e5384487a770b223961b9ed6147aa70f30b1718460e76d17160c1183" Netns:"/var/run/netns/03eb3f0d-19a7-47d6-b9da-494001977c66" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-7d9549f6c-gj4pr;K8S_POD_INFRA_CONTAINER_ID=1fc32ef2e5384487a770b223961b9ed6147aa70f30b1718460e76d17160c1183;K8S_POD_UID=8ab5a672-80db-49a0-9b83-db5b7a5b4e65" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-7d9549f6c-gj4pr] networking: Multus: [openshift-authentication/oauth-openshift-7d9549f6c-gj4pr/8ab5a672-80db-49a0-9b83-db5b7a5b4e65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod oauth-openshift-7d9549f6c-gj4pr in out of cluster comm: SetNetworkStatus: failed to update the pod oauth-openshift-7d9549f6c-gj4pr in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7d9549f6c-gj4pr?timeout=1m0s": dial tcp 38.102.83.214:6443: connect: connection refused Dec 10 18:59:24 crc kubenswrapper[4828]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 10 18:59:24 crc kubenswrapper[4828]: > Dec 10 18:59:24 crc kubenswrapper[4828]: E1210 18:59:24.881276 4828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 10 18:59:24 crc kubenswrapper[4828]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-7d9549f6c-gj4pr_openshift-authentication_8ab5a672-80db-49a0-9b83-db5b7a5b4e65_0(1fc32ef2e5384487a770b223961b9ed6147aa70f30b1718460e76d17160c1183): error adding pod openshift-authentication_oauth-openshift-7d9549f6c-gj4pr to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"1fc32ef2e5384487a770b223961b9ed6147aa70f30b1718460e76d17160c1183" Netns:"/var/run/netns/03eb3f0d-19a7-47d6-b9da-494001977c66" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-7d9549f6c-gj4pr;K8S_POD_INFRA_CONTAINER_ID=1fc32ef2e5384487a770b223961b9ed6147aa70f30b1718460e76d17160c1183;K8S_POD_UID=8ab5a672-80db-49a0-9b83-db5b7a5b4e65" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-7d9549f6c-gj4pr] networking: Multus: [openshift-authentication/oauth-openshift-7d9549f6c-gj4pr/8ab5a672-80db-49a0-9b83-db5b7a5b4e65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod oauth-openshift-7d9549f6c-gj4pr in out of cluster comm: SetNetworkStatus: failed to update the pod oauth-openshift-7d9549f6c-gj4pr in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7d9549f6c-gj4pr?timeout=1m0s": dial tcp 38.102.83.214:6443: connect: connection refused Dec 10 18:59:24 crc kubenswrapper[4828]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 10 18:59:24 crc kubenswrapper[4828]: > pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:24 crc kubenswrapper[4828]: E1210 18:59:24.881309 4828 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 10 18:59:24 crc kubenswrapper[4828]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-7d9549f6c-gj4pr_openshift-authentication_8ab5a672-80db-49a0-9b83-db5b7a5b4e65_0(1fc32ef2e5384487a770b223961b9ed6147aa70f30b1718460e76d17160c1183): error adding pod openshift-authentication_oauth-openshift-7d9549f6c-gj4pr to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"1fc32ef2e5384487a770b223961b9ed6147aa70f30b1718460e76d17160c1183" Netns:"/var/run/netns/03eb3f0d-19a7-47d6-b9da-494001977c66" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-7d9549f6c-gj4pr;K8S_POD_INFRA_CONTAINER_ID=1fc32ef2e5384487a770b223961b9ed6147aa70f30b1718460e76d17160c1183;K8S_POD_UID=8ab5a672-80db-49a0-9b83-db5b7a5b4e65" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-7d9549f6c-gj4pr] networking: Multus: [openshift-authentication/oauth-openshift-7d9549f6c-gj4pr/8ab5a672-80db-49a0-9b83-db5b7a5b4e65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod oauth-openshift-7d9549f6c-gj4pr in out of cluster comm: SetNetworkStatus: failed to update the pod oauth-openshift-7d9549f6c-gj4pr in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7d9549f6c-gj4pr?timeout=1m0s": dial tcp 38.102.83.214:6443: connect: connection refused Dec 10 18:59:24 crc kubenswrapper[4828]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 10 18:59:24 crc kubenswrapper[4828]: > pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:24 crc kubenswrapper[4828]: E1210 18:59:24.881420 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"oauth-openshift-7d9549f6c-gj4pr_openshift-authentication(8ab5a672-80db-49a0-9b83-db5b7a5b4e65)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"oauth-openshift-7d9549f6c-gj4pr_openshift-authentication(8ab5a672-80db-49a0-9b83-db5b7a5b4e65)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-7d9549f6c-gj4pr_openshift-authentication_8ab5a672-80db-49a0-9b83-db5b7a5b4e65_0(1fc32ef2e5384487a770b223961b9ed6147aa70f30b1718460e76d17160c1183): error adding pod openshift-authentication_oauth-openshift-7d9549f6c-gj4pr to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"1fc32ef2e5384487a770b223961b9ed6147aa70f30b1718460e76d17160c1183\\\" Netns:\\\"/var/run/netns/03eb3f0d-19a7-47d6-b9da-494001977c66\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-7d9549f6c-gj4pr;K8S_POD_INFRA_CONTAINER_ID=1fc32ef2e5384487a770b223961b9ed6147aa70f30b1718460e76d17160c1183;K8S_POD_UID=8ab5a672-80db-49a0-9b83-db5b7a5b4e65\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-7d9549f6c-gj4pr] networking: Multus: [openshift-authentication/oauth-openshift-7d9549f6c-gj4pr/8ab5a672-80db-49a0-9b83-db5b7a5b4e65]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod oauth-openshift-7d9549f6c-gj4pr in out of cluster comm: SetNetworkStatus: failed to update the pod oauth-openshift-7d9549f6c-gj4pr in out of cluster comm: status update failed for pod /: Get \\\"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-7d9549f6c-gj4pr?timeout=1m0s\\\": dial tcp 38.102.83.214:6443: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" podUID="8ab5a672-80db-49a0-9b83-db5b7a5b4e65" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.448993 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.449046 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1b642aaded5acb453d010c103be1d2fcfbf97f57a8ee3a5d452d0621cec1ba29"} Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.450323 4828 status_manager.go:851] "Failed to get status for pod" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" pod="openshift-marketplace/redhat-marketplace-ftr4x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ftr4x\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.450826 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.451242 4828 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.451551 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.451842 4828 status_manager.go:851] "Failed to get status for pod" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" pod="openshift-marketplace/redhat-marketplace-qs5qt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-qs5qt\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.452065 4828 status_manager.go:851] "Failed to get status for pod" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" pod="openshift-marketplace/redhat-operators-5d8xc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5d8xc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.452325 4828 status_manager.go:851] "Failed to get status for pod" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" pod="openshift-marketplace/redhat-operators-kfgtf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kfgtf\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.452612 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.452845 4828 status_manager.go:851] "Failed to get status for pod" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" pod="openshift-marketplace/certified-operators-2vvgb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2vvgb\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.453101 4828 status_manager.go:851] "Failed to get status for pod" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" pod="openshift-marketplace/certified-operators-gs2g5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gs2g5\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.453384 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.453608 4828 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.763186 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qs5qt" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.763224 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qs5qt" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.788395 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.789288 4828 status_manager.go:851] "Failed to get status for pod" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" pod="openshift-marketplace/certified-operators-2vvgb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2vvgb\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.789637 4828 status_manager.go:851] "Failed to get status for pod" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" pod="openshift-marketplace/certified-operators-gs2g5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gs2g5\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.789849 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.790013 4828 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.790149 4828 status_manager.go:851] "Failed to get status for pod" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" pod="openshift-marketplace/redhat-marketplace-ftr4x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ftr4x\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.790280 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.790413 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.790557 4828 status_manager.go:851] "Failed to get status for pod" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" pod="openshift-marketplace/redhat-marketplace-qs5qt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-qs5qt\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.790689 4828 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.792347 4828 status_manager.go:851] "Failed to get status for pod" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" pod="openshift-marketplace/redhat-operators-5d8xc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5d8xc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.792737 4828 status_manager.go:851] "Failed to get status for pod" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" pod="openshift-marketplace/redhat-operators-kfgtf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kfgtf\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.793153 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.805173 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qs5qt" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.805571 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.805744 4828 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.805976 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.806296 4828 status_manager.go:851] "Failed to get status for pod" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" pod="openshift-marketplace/redhat-marketplace-qs5qt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-qs5qt\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.806703 4828 status_manager.go:851] "Failed to get status for pod" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" pod="openshift-marketplace/redhat-operators-5d8xc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5d8xc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.806948 4828 status_manager.go:851] "Failed to get status for pod" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" pod="openshift-marketplace/redhat-operators-kfgtf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kfgtf\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.807211 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.807428 4828 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fe272100-65b7-4be3-9ab4-51e4a6f475de" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.807448 4828 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fe272100-65b7-4be3-9ab4-51e4a6f475de" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.807594 4828 status_manager.go:851] "Failed to get status for pod" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" pod="openshift-marketplace/certified-operators-2vvgb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2vvgb\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: E1210 18:59:25.807677 4828 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.807999 4828 status_manager.go:851] "Failed to get status for pod" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" pod="openshift-marketplace/certified-operators-gs2g5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gs2g5\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.808164 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.808190 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:25 crc kubenswrapper[4828]: I1210 18:59:25.808372 4828 status_manager.go:851] "Failed to get status for pod" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" pod="openshift-marketplace/redhat-marketplace-ftr4x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ftr4x\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.076781 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.141789 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ftr4x" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.141932 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ftr4x" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.186915 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ftr4x" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.187453 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.187810 4828 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.187979 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.188128 4828 status_manager.go:851] "Failed to get status for pod" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" pod="openshift-marketplace/redhat-marketplace-qs5qt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-qs5qt\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.188285 4828 status_manager.go:851] "Failed to get status for pod" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" pod="openshift-marketplace/redhat-operators-5d8xc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5d8xc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.188427 4828 status_manager.go:851] "Failed to get status for pod" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" pod="openshift-marketplace/redhat-operators-kfgtf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kfgtf\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.188569 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.188706 4828 status_manager.go:851] "Failed to get status for pod" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" pod="openshift-marketplace/certified-operators-2vvgb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2vvgb\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.188909 4828 status_manager.go:851] "Failed to get status for pod" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" pod="openshift-marketplace/certified-operators-gs2g5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gs2g5\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.189058 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.189202 4828 status_manager.go:851] "Failed to get status for pod" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" pod="openshift-marketplace/redhat-marketplace-ftr4x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ftr4x\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.456645 4828 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="0eb0972094c371b7e862ebc5407dd9c099c1b3838d5db4c0df5153135d2e4e93" exitCode=0 Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.457463 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"0eb0972094c371b7e862ebc5407dd9c099c1b3838d5db4c0df5153135d2e4e93"} Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.457500 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e62ec0ff3220d66bb126e24e14fac9019e90fd23c289e0ddf10468464cbc13a6"} Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.457702 4828 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fe272100-65b7-4be3-9ab4-51e4a6f475de" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.457720 4828 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fe272100-65b7-4be3-9ab4-51e4a6f475de" Dec 10 18:59:26 crc kubenswrapper[4828]: E1210 18:59:26.458157 4828 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.458386 4828 status_manager.go:851] "Failed to get status for pod" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" pod="openshift-marketplace/certified-operators-2vvgb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-2vvgb\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.458778 4828 status_manager.go:851] "Failed to get status for pod" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" pod="openshift-marketplace/certified-operators-gs2g5" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-gs2g5\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.459286 4828 status_manager.go:851] "Failed to get status for pod" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" pod="openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-5b5bf74995-db8pv\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.459662 4828 status_manager.go:851] "Failed to get status for pod" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" pod="openshift-marketplace/redhat-marketplace-ftr4x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-ftr4x\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.460083 4828 status_manager.go:851] "Failed to get status for pod" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.460745 4828 status_manager.go:851] "Failed to get status for pod" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" pod="openshift-marketplace/redhat-marketplace-qs5qt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-qs5qt\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.461113 4828 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.461401 4828 status_manager.go:851] "Failed to get status for pod" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" pod="openshift-marketplace/community-operators-ll5gw" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-ll5gw\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.461668 4828 status_manager.go:851] "Failed to get status for pod" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" pod="openshift-marketplace/redhat-operators-kfgtf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-kfgtf\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.461939 4828 status_manager.go:851] "Failed to get status for pod" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" pod="openshift-marketplace/redhat-operators-5d8xc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-5d8xc\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:26 crc kubenswrapper[4828]: I1210 18:59:26.462301 4828 status_manager.go:851] "Failed to get status for pod" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" pod="openshift-controller-manager/controller-manager-56f99f9589-4ljp7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-56f99f9589-4ljp7\": dial tcp 38.102.83.214:6443: connect: connection refused" Dec 10 18:59:27 crc kubenswrapper[4828]: I1210 18:59:27.464887 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7c97ffaf9ab4ab7729251ebeb355d547b250262e9e74065fc08e7227f21d690f"} Dec 10 18:59:27 crc kubenswrapper[4828]: I1210 18:59:27.465215 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"bc07feefb53e887c16de45c105f0d19adead994b0306825aa47ad3ba73b60aab"} Dec 10 18:59:27 crc kubenswrapper[4828]: I1210 18:59:27.465227 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a5cce5fcf296169d3ccdd74fad9edd57798ca6eb38c9e862fa17e7e4834142c4"} Dec 10 18:59:27 crc kubenswrapper[4828]: I1210 18:59:27.465235 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e2025e3e19594bc46f0a45f9333d16b9df41d12239565c149e96637cfe940bb6"} Dec 10 18:59:27 crc kubenswrapper[4828]: I1210 18:59:27.472248 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5d8xc" Dec 10 18:59:27 crc kubenswrapper[4828]: I1210 18:59:27.472310 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5d8xc" Dec 10 18:59:27 crc kubenswrapper[4828]: I1210 18:59:27.549517 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kfgtf" Dec 10 18:59:27 crc kubenswrapper[4828]: I1210 18:59:27.549567 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kfgtf" Dec 10 18:59:28 crc kubenswrapper[4828]: I1210 18:59:28.475352 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"18e176a0213e64a38fe8ca7929e1ba0a85580d89a7b3979d1ed34d0629a91bff"} Dec 10 18:59:28 crc kubenswrapper[4828]: I1210 18:59:28.475544 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:28 crc kubenswrapper[4828]: I1210 18:59:28.475668 4828 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fe272100-65b7-4be3-9ab4-51e4a6f475de" Dec 10 18:59:28 crc kubenswrapper[4828]: I1210 18:59:28.475698 4828 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fe272100-65b7-4be3-9ab4-51e4a6f475de" Dec 10 18:59:28 crc kubenswrapper[4828]: I1210 18:59:28.507959 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5d8xc" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" containerName="registry-server" probeResult="failure" output=< Dec 10 18:59:28 crc kubenswrapper[4828]: timeout: failed to connect service ":50051" within 1s Dec 10 18:59:28 crc kubenswrapper[4828]: > Dec 10 18:59:28 crc kubenswrapper[4828]: I1210 18:59:28.586455 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kfgtf" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" containerName="registry-server" probeResult="failure" output=< Dec 10 18:59:28 crc kubenswrapper[4828]: timeout: failed to connect service ":50051" within 1s Dec 10 18:59:28 crc kubenswrapper[4828]: > Dec 10 18:59:30 crc kubenswrapper[4828]: I1210 18:59:30.809083 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:30 crc kubenswrapper[4828]: I1210 18:59:30.809187 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:30 crc kubenswrapper[4828]: I1210 18:59:30.816684 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:33 crc kubenswrapper[4828]: I1210 18:59:33.487737 4828 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:33 crc kubenswrapper[4828]: I1210 18:59:33.725429 4828 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8188e72d-28c8-4c31-b65b-10f5d37bb1a6" Dec 10 18:59:33 crc kubenswrapper[4828]: I1210 18:59:33.781690 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gs2g5" Dec 10 18:59:33 crc kubenswrapper[4828]: I1210 18:59:33.781753 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gs2g5" Dec 10 18:59:33 crc kubenswrapper[4828]: I1210 18:59:33.807748 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:59:33 crc kubenswrapper[4828]: I1210 18:59:33.811361 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:59:33 crc kubenswrapper[4828]: I1210 18:59:33.820340 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gs2g5" Dec 10 18:59:34 crc kubenswrapper[4828]: I1210 18:59:34.192897 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2vvgb" Dec 10 18:59:34 crc kubenswrapper[4828]: I1210 18:59:34.192967 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2vvgb" Dec 10 18:59:34 crc kubenswrapper[4828]: I1210 18:59:34.236786 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2vvgb" Dec 10 18:59:34 crc kubenswrapper[4828]: I1210 18:59:34.506372 4828 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fe272100-65b7-4be3-9ab4-51e4a6f475de" Dec 10 18:59:34 crc kubenswrapper[4828]: I1210 18:59:34.506411 4828 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fe272100-65b7-4be3-9ab4-51e4a6f475de" Dec 10 18:59:34 crc kubenswrapper[4828]: I1210 18:59:34.510155 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 10 18:59:34 crc kubenswrapper[4828]: I1210 18:59:34.510856 4828 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8188e72d-28c8-4c31-b65b-10f5d37bb1a6" Dec 10 18:59:34 crc kubenswrapper[4828]: I1210 18:59:34.562596 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gs2g5" Dec 10 18:59:34 crc kubenswrapper[4828]: I1210 18:59:34.571945 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2vvgb" Dec 10 18:59:35 crc kubenswrapper[4828]: I1210 18:59:35.824351 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qs5qt" Dec 10 18:59:36 crc kubenswrapper[4828]: I1210 18:59:36.195913 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ftr4x" Dec 10 18:59:36 crc kubenswrapper[4828]: I1210 18:59:36.787881 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:36 crc kubenswrapper[4828]: I1210 18:59:36.788374 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:37 crc kubenswrapper[4828]: W1210 18:59:37.217243 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ab5a672_80db_49a0_9b83_db5b7a5b4e65.slice/crio-18cc849e070bd897239695f2202a09caaa7f32ec8d4ce3dda869c12d444e2c21 WatchSource:0}: Error finding container 18cc849e070bd897239695f2202a09caaa7f32ec8d4ce3dda869c12d444e2c21: Status 404 returned error can't find the container with id 18cc849e070bd897239695f2202a09caaa7f32ec8d4ce3dda869c12d444e2c21 Dec 10 18:59:37 crc kubenswrapper[4828]: I1210 18:59:37.523364 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5d8xc" Dec 10 18:59:37 crc kubenswrapper[4828]: I1210 18:59:37.525478 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" event={"ID":"8ab5a672-80db-49a0-9b83-db5b7a5b4e65","Type":"ContainerStarted","Data":"18cc849e070bd897239695f2202a09caaa7f32ec8d4ce3dda869c12d444e2c21"} Dec 10 18:59:37 crc kubenswrapper[4828]: I1210 18:59:37.563165 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5d8xc" Dec 10 18:59:37 crc kubenswrapper[4828]: I1210 18:59:37.586762 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kfgtf" Dec 10 18:59:37 crc kubenswrapper[4828]: I1210 18:59:37.621198 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kfgtf" Dec 10 18:59:38 crc kubenswrapper[4828]: I1210 18:59:38.534506 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" event={"ID":"8ab5a672-80db-49a0-9b83-db5b7a5b4e65","Type":"ContainerStarted","Data":"eae3136749bfcc9d09d0233f04dc456e829ac139b0c1cc5d3ca6ebbcea7dcf50"} Dec 10 18:59:38 crc kubenswrapper[4828]: I1210 18:59:38.535389 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:38 crc kubenswrapper[4828]: I1210 18:59:38.554473 4828 patch_prober.go:28] interesting pod/oauth-openshift-7d9549f6c-gj4pr container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.58:6443/healthz\": read tcp 10.217.0.2:40416->10.217.0.58:6443: read: connection reset by peer" start-of-body= Dec 10 18:59:38 crc kubenswrapper[4828]: I1210 18:59:38.554529 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" podUID="8ab5a672-80db-49a0-9b83-db5b7a5b4e65" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.58:6443/healthz\": read tcp 10.217.0.2:40416->10.217.0.58:6443: read: connection reset by peer" Dec 10 18:59:39 crc kubenswrapper[4828]: I1210 18:59:39.541839 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7d9549f6c-gj4pr_8ab5a672-80db-49a0-9b83-db5b7a5b4e65/oauth-openshift/0.log" Dec 10 18:59:39 crc kubenswrapper[4828]: I1210 18:59:39.541914 4828 generic.go:334] "Generic (PLEG): container finished" podID="8ab5a672-80db-49a0-9b83-db5b7a5b4e65" containerID="eae3136749bfcc9d09d0233f04dc456e829ac139b0c1cc5d3ca6ebbcea7dcf50" exitCode=255 Dec 10 18:59:39 crc kubenswrapper[4828]: I1210 18:59:39.541954 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" event={"ID":"8ab5a672-80db-49a0-9b83-db5b7a5b4e65","Type":"ContainerDied","Data":"eae3136749bfcc9d09d0233f04dc456e829ac139b0c1cc5d3ca6ebbcea7dcf50"} Dec 10 18:59:39 crc kubenswrapper[4828]: I1210 18:59:39.542761 4828 scope.go:117] "RemoveContainer" containerID="eae3136749bfcc9d09d0233f04dc456e829ac139b0c1cc5d3ca6ebbcea7dcf50" Dec 10 18:59:40 crc kubenswrapper[4828]: I1210 18:59:40.550742 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7d9549f6c-gj4pr_8ab5a672-80db-49a0-9b83-db5b7a5b4e65/oauth-openshift/1.log" Dec 10 18:59:40 crc kubenswrapper[4828]: I1210 18:59:40.551525 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7d9549f6c-gj4pr_8ab5a672-80db-49a0-9b83-db5b7a5b4e65/oauth-openshift/0.log" Dec 10 18:59:40 crc kubenswrapper[4828]: I1210 18:59:40.551599 4828 generic.go:334] "Generic (PLEG): container finished" podID="8ab5a672-80db-49a0-9b83-db5b7a5b4e65" containerID="109b9cc57062b7758b719816a1707eb10ea6545c9865b1babb0d2b0fb3dc4048" exitCode=255 Dec 10 18:59:40 crc kubenswrapper[4828]: I1210 18:59:40.551640 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" event={"ID":"8ab5a672-80db-49a0-9b83-db5b7a5b4e65","Type":"ContainerDied","Data":"109b9cc57062b7758b719816a1707eb10ea6545c9865b1babb0d2b0fb3dc4048"} Dec 10 18:59:40 crc kubenswrapper[4828]: I1210 18:59:40.551682 4828 scope.go:117] "RemoveContainer" containerID="eae3136749bfcc9d09d0233f04dc456e829ac139b0c1cc5d3ca6ebbcea7dcf50" Dec 10 18:59:40 crc kubenswrapper[4828]: I1210 18:59:40.552296 4828 scope.go:117] "RemoveContainer" containerID="109b9cc57062b7758b719816a1707eb10ea6545c9865b1babb0d2b0fb3dc4048" Dec 10 18:59:40 crc kubenswrapper[4828]: E1210 18:59:40.552623 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-7d9549f6c-gj4pr_openshift-authentication(8ab5a672-80db-49a0-9b83-db5b7a5b4e65)\"" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" podUID="8ab5a672-80db-49a0-9b83-db5b7a5b4e65" Dec 10 18:59:41 crc kubenswrapper[4828]: I1210 18:59:41.558120 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7d9549f6c-gj4pr_8ab5a672-80db-49a0-9b83-db5b7a5b4e65/oauth-openshift/1.log" Dec 10 18:59:41 crc kubenswrapper[4828]: I1210 18:59:41.558986 4828 scope.go:117] "RemoveContainer" containerID="109b9cc57062b7758b719816a1707eb10ea6545c9865b1babb0d2b0fb3dc4048" Dec 10 18:59:41 crc kubenswrapper[4828]: E1210 18:59:41.559372 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-7d9549f6c-gj4pr_openshift-authentication(8ab5a672-80db-49a0-9b83-db5b7a5b4e65)\"" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" podUID="8ab5a672-80db-49a0-9b83-db5b7a5b4e65" Dec 10 18:59:42 crc kubenswrapper[4828]: I1210 18:59:42.685692 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 10 18:59:43 crc kubenswrapper[4828]: I1210 18:59:43.346267 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 10 18:59:43 crc kubenswrapper[4828]: I1210 18:59:43.478235 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 10 18:59:43 crc kubenswrapper[4828]: I1210 18:59:43.583793 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:43 crc kubenswrapper[4828]: I1210 18:59:43.583883 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:43 crc kubenswrapper[4828]: I1210 18:59:43.584547 4828 scope.go:117] "RemoveContainer" containerID="109b9cc57062b7758b719816a1707eb10ea6545c9865b1babb0d2b0fb3dc4048" Dec 10 18:59:43 crc kubenswrapper[4828]: E1210 18:59:43.584926 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-7d9549f6c-gj4pr_openshift-authentication(8ab5a672-80db-49a0-9b83-db5b7a5b4e65)\"" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" podUID="8ab5a672-80db-49a0-9b83-db5b7a5b4e65" Dec 10 18:59:43 crc kubenswrapper[4828]: I1210 18:59:43.839031 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 10 18:59:43 crc kubenswrapper[4828]: I1210 18:59:43.965932 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 10 18:59:44 crc kubenswrapper[4828]: I1210 18:59:44.040419 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 10 18:59:44 crc kubenswrapper[4828]: I1210 18:59:44.392553 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 10 18:59:44 crc kubenswrapper[4828]: I1210 18:59:44.415201 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 10 18:59:44 crc kubenswrapper[4828]: I1210 18:59:44.524900 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 10 18:59:45 crc kubenswrapper[4828]: I1210 18:59:45.046929 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 10 18:59:45 crc kubenswrapper[4828]: I1210 18:59:45.187019 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 10 18:59:45 crc kubenswrapper[4828]: I1210 18:59:45.202119 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 10 18:59:45 crc kubenswrapper[4828]: I1210 18:59:45.386926 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 10 18:59:45 crc kubenswrapper[4828]: I1210 18:59:45.436616 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 10 18:59:45 crc kubenswrapper[4828]: I1210 18:59:45.514336 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 10 18:59:45 crc kubenswrapper[4828]: I1210 18:59:45.538762 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 10 18:59:45 crc kubenswrapper[4828]: I1210 18:59:45.551263 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 10 18:59:45 crc kubenswrapper[4828]: I1210 18:59:45.718039 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 10 18:59:45 crc kubenswrapper[4828]: I1210 18:59:45.858065 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 10 18:59:45 crc kubenswrapper[4828]: I1210 18:59:45.888583 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 10 18:59:45 crc kubenswrapper[4828]: I1210 18:59:45.906535 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 10 18:59:45 crc kubenswrapper[4828]: I1210 18:59:45.922841 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.069642 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.157202 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.194941 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.226685 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.470248 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.625552 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.831050 4828 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.831656 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gs2g5" podStartSLOduration=25.705638991 podStartE2EDuration="2m3.831636136s" podCreationTimestamp="2025-12-10 18:57:43 +0000 UTC" firstStartedPulling="2025-12-10 18:57:45.508523665 +0000 UTC m=+146.019134670" lastFinishedPulling="2025-12-10 18:59:23.63452081 +0000 UTC m=+244.145131815" observedRunningTime="2025-12-10 18:59:33.690285825 +0000 UTC m=+254.200896820" watchObservedRunningTime="2025-12-10 18:59:46.831636136 +0000 UTC m=+267.342247141" Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.832360 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qs5qt" podStartSLOduration=24.789676542 podStartE2EDuration="2m1.832354306s" podCreationTimestamp="2025-12-10 18:57:45 +0000 UTC" firstStartedPulling="2025-12-10 18:57:46.524951964 +0000 UTC m=+147.035562969" lastFinishedPulling="2025-12-10 18:59:23.567629728 +0000 UTC m=+244.078240733" observedRunningTime="2025-12-10 18:59:33.594773068 +0000 UTC m=+254.105384083" watchObservedRunningTime="2025-12-10 18:59:46.832354306 +0000 UTC m=+267.342965311" Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.833682 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ftr4x" podStartSLOduration=24.781186013 podStartE2EDuration="2m1.833676262s" podCreationTimestamp="2025-12-10 18:57:45 +0000 UTC" firstStartedPulling="2025-12-10 18:57:46.528651995 +0000 UTC m=+147.039263000" lastFinishedPulling="2025-12-10 18:59:23.581142234 +0000 UTC m=+244.091753249" observedRunningTime="2025-12-10 18:59:33.509479617 +0000 UTC m=+254.020090622" watchObservedRunningTime="2025-12-10 18:59:46.833676262 +0000 UTC m=+267.344287267" Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.833786 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kfgtf" podStartSLOduration=25.837186634 podStartE2EDuration="1m59.833781765s" podCreationTimestamp="2025-12-10 18:57:47 +0000 UTC" firstStartedPulling="2025-12-10 18:57:49.662298029 +0000 UTC m=+150.172909034" lastFinishedPulling="2025-12-10 18:59:23.65889316 +0000 UTC m=+244.169504165" observedRunningTime="2025-12-10 18:59:33.626033365 +0000 UTC m=+254.136644370" watchObservedRunningTime="2025-12-10 18:59:46.833781765 +0000 UTC m=+267.344392780" Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.834771 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5d8xc" podStartSLOduration=30.873014182 podStartE2EDuration="2m0.834764622s" podCreationTimestamp="2025-12-10 18:57:46 +0000 UTC" firstStartedPulling="2025-12-10 18:57:49.666159492 +0000 UTC m=+150.176770497" lastFinishedPulling="2025-12-10 18:59:19.627909932 +0000 UTC m=+240.138520937" observedRunningTime="2025-12-10 18:59:33.610437132 +0000 UTC m=+254.121048137" watchObservedRunningTime="2025-12-10 18:59:46.834764622 +0000 UTC m=+267.345375627" Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.835275 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2vvgb" podStartSLOduration=25.706274491 podStartE2EDuration="2m3.835271166s" podCreationTimestamp="2025-12-10 18:57:43 +0000 UTC" firstStartedPulling="2025-12-10 18:57:45.506533102 +0000 UTC m=+146.017144107" lastFinishedPulling="2025-12-10 18:59:23.635529777 +0000 UTC m=+244.146140782" observedRunningTime="2025-12-10 18:59:33.658630958 +0000 UTC m=+254.169241963" watchObservedRunningTime="2025-12-10 18:59:46.835271166 +0000 UTC m=+267.345882171" Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.835610 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-controller-manager/controller-manager-56f99f9589-4ljp7","openshift-route-controller-manager/route-controller-manager-5b5bf74995-db8pv"] Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.835662 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.835694 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d9549f6c-gj4pr"] Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.836350 4828 scope.go:117] "RemoveContainer" containerID="109b9cc57062b7758b719816a1707eb10ea6545c9865b1babb0d2b0fb3dc4048" Dec 10 18:59:46 crc kubenswrapper[4828]: E1210 18:59:46.836539 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-7d9549f6c-gj4pr_openshift-authentication(8ab5a672-80db-49a0-9b83-db5b7a5b4e65)\"" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" podUID="8ab5a672-80db-49a0-9b83-db5b7a5b4e65" Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.855296 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=13.855277898 podStartE2EDuration="13.855277898s" podCreationTimestamp="2025-12-10 18:59:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:59:46.8510027 +0000 UTC m=+267.361613715" watchObservedRunningTime="2025-12-10 18:59:46.855277898 +0000 UTC m=+267.365888903" Dec 10 18:59:46 crc kubenswrapper[4828]: I1210 18:59:46.927832 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.014259 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.054275 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.054311 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.059444 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.114421 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.135343 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.140951 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.161873 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.206356 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.212507 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.260264 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.270255 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.287360 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.292251 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.293849 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.566554 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.679593 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.683125 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.703911 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.704773 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.713006 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.782135 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.795682 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" path="/var/lib/kubelet/pods/34219f30-45e9-4be7-9b99-ddc8046fccbe/volumes" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.821886 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" path="/var/lib/kubelet/pods/85621df9-138d-4c2e-a734-74dd69d9195e/volumes" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.962375 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.963512 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 10 18:59:47 crc kubenswrapper[4828]: I1210 18:59:47.992233 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 10 18:59:48 crc kubenswrapper[4828]: I1210 18:59:48.151053 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 10 18:59:48 crc kubenswrapper[4828]: I1210 18:59:48.295334 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 10 18:59:48 crc kubenswrapper[4828]: I1210 18:59:48.368432 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 10 18:59:48 crc kubenswrapper[4828]: I1210 18:59:48.408020 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 10 18:59:48 crc kubenswrapper[4828]: I1210 18:59:48.438954 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 10 18:59:48 crc kubenswrapper[4828]: I1210 18:59:48.505015 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 10 18:59:48 crc kubenswrapper[4828]: I1210 18:59:48.507240 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 10 18:59:48 crc kubenswrapper[4828]: I1210 18:59:48.516183 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 10 18:59:48 crc kubenswrapper[4828]: I1210 18:59:48.556485 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 10 18:59:48 crc kubenswrapper[4828]: I1210 18:59:48.750044 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 10 18:59:48 crc kubenswrapper[4828]: I1210 18:59:48.787924 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 10 18:59:48 crc kubenswrapper[4828]: I1210 18:59:48.846351 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 10 18:59:48 crc kubenswrapper[4828]: I1210 18:59:48.931073 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.010833 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.110740 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.120881 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.122706 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.209263 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.263822 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.349170 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.392225 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.467020 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.556334 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.577461 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.596467 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.612987 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.625904 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.688008 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.709001 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.760152 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.797763 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.862990 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 10 18:59:49 crc kubenswrapper[4828]: I1210 18:59:49.988604 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 10 18:59:50 crc kubenswrapper[4828]: I1210 18:59:50.029570 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 10 18:59:50 crc kubenswrapper[4828]: I1210 18:59:50.126047 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 10 18:59:50 crc kubenswrapper[4828]: I1210 18:59:50.212558 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 10 18:59:50 crc kubenswrapper[4828]: I1210 18:59:50.403683 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 10 18:59:50 crc kubenswrapper[4828]: I1210 18:59:50.423895 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 10 18:59:50 crc kubenswrapper[4828]: I1210 18:59:50.429025 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 10 18:59:50 crc kubenswrapper[4828]: I1210 18:59:50.498444 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 10 18:59:50 crc kubenswrapper[4828]: I1210 18:59:50.522538 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 10 18:59:50 crc kubenswrapper[4828]: I1210 18:59:50.552665 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 10 18:59:50 crc kubenswrapper[4828]: I1210 18:59:50.714882 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 10 18:59:50 crc kubenswrapper[4828]: I1210 18:59:50.719995 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 10 18:59:50 crc kubenswrapper[4828]: I1210 18:59:50.726457 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 10 18:59:50 crc kubenswrapper[4828]: I1210 18:59:50.812772 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:50 crc kubenswrapper[4828]: I1210 18:59:50.817075 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 10 18:59:50 crc kubenswrapper[4828]: I1210 18:59:50.854918 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 10 18:59:50 crc kubenswrapper[4828]: I1210 18:59:50.987155 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 10 18:59:51 crc kubenswrapper[4828]: I1210 18:59:51.053256 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 10 18:59:51 crc kubenswrapper[4828]: I1210 18:59:51.127949 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 10 18:59:51 crc kubenswrapper[4828]: I1210 18:59:51.230858 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 10 18:59:51 crc kubenswrapper[4828]: I1210 18:59:51.231653 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 10 18:59:51 crc kubenswrapper[4828]: I1210 18:59:51.248760 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 10 18:59:51 crc kubenswrapper[4828]: I1210 18:59:51.270881 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 10 18:59:51 crc kubenswrapper[4828]: I1210 18:59:51.450919 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 10 18:59:51 crc kubenswrapper[4828]: I1210 18:59:51.488076 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 10 18:59:51 crc kubenswrapper[4828]: I1210 18:59:51.496542 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 10 18:59:51 crc kubenswrapper[4828]: I1210 18:59:51.529167 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 10 18:59:51 crc kubenswrapper[4828]: I1210 18:59:51.686190 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 10 18:59:51 crc kubenswrapper[4828]: I1210 18:59:51.878304 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 10 18:59:51 crc kubenswrapper[4828]: I1210 18:59:51.883556 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.056211 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.091137 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.104739 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.111058 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.120679 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.211825 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.212978 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.309986 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.313517 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.345082 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.367624 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.393327 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.591090 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.616116 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.672232 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.677515 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.700298 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.861361 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.926708 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.944826 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 10 18:59:52 crc kubenswrapper[4828]: I1210 18:59:52.976395 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.009915 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.069958 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.194305 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.222316 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.240838 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t"] Dec 10 18:59:53 crc kubenswrapper[4828]: E1210 18:59:53.241111 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" containerName="controller-manager" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.241131 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" containerName="controller-manager" Dec 10 18:59:53 crc kubenswrapper[4828]: E1210 18:59:53.241142 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" containerName="installer" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.241148 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" containerName="installer" Dec 10 18:59:53 crc kubenswrapper[4828]: E1210 18:59:53.241166 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" containerName="route-controller-manager" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.241172 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" containerName="route-controller-manager" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.241271 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="867406bf-a6e6-4644-a18b-4fa778cea1a5" containerName="installer" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.241285 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="34219f30-45e9-4be7-9b99-ddc8046fccbe" containerName="controller-manager" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.241294 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="85621df9-138d-4c2e-a734-74dd69d9195e" containerName="route-controller-manager" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.241702 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.243263 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.244487 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f"] Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.245168 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.246878 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.247209 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.247596 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.247789 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.248464 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.248481 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.248618 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.249374 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.249643 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.250386 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.250458 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.250549 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.252114 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t"] Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.254668 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.261014 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f"] Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.370609 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1-proxy-ca-bundles\") pod \"controller-manager-7f5d9b8d98-pwz4t\" (UID: \"b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1\") " pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.370661 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1-client-ca\") pod \"controller-manager-7f5d9b8d98-pwz4t\" (UID: \"b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1\") " pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.370691 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f713c391-bba5-4d61-a342-bebc6757fd97-serving-cert\") pod \"route-controller-manager-5b6994579d-4mj7f\" (UID: \"f713c391-bba5-4d61-a342-bebc6757fd97\") " pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.370774 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1-config\") pod \"controller-manager-7f5d9b8d98-pwz4t\" (UID: \"b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1\") " pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.370848 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1-serving-cert\") pod \"controller-manager-7f5d9b8d98-pwz4t\" (UID: \"b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1\") " pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.370947 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf8sq\" (UniqueName: \"kubernetes.io/projected/b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1-kube-api-access-lf8sq\") pod \"controller-manager-7f5d9b8d98-pwz4t\" (UID: \"b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1\") " pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.371001 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f713c391-bba5-4d61-a342-bebc6757fd97-client-ca\") pod \"route-controller-manager-5b6994579d-4mj7f\" (UID: \"f713c391-bba5-4d61-a342-bebc6757fd97\") " pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.371023 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f713c391-bba5-4d61-a342-bebc6757fd97-config\") pod \"route-controller-manager-5b6994579d-4mj7f\" (UID: \"f713c391-bba5-4d61-a342-bebc6757fd97\") " pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.371047 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lsrq\" (UniqueName: \"kubernetes.io/projected/f713c391-bba5-4d61-a342-bebc6757fd97-kube-api-access-9lsrq\") pod \"route-controller-manager-5b6994579d-4mj7f\" (UID: \"f713c391-bba5-4d61-a342-bebc6757fd97\") " pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.395749 4828 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.447928 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.471648 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f713c391-bba5-4d61-a342-bebc6757fd97-client-ca\") pod \"route-controller-manager-5b6994579d-4mj7f\" (UID: \"f713c391-bba5-4d61-a342-bebc6757fd97\") " pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.471945 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f713c391-bba5-4d61-a342-bebc6757fd97-config\") pod \"route-controller-manager-5b6994579d-4mj7f\" (UID: \"f713c391-bba5-4d61-a342-bebc6757fd97\") " pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.472044 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lsrq\" (UniqueName: \"kubernetes.io/projected/f713c391-bba5-4d61-a342-bebc6757fd97-kube-api-access-9lsrq\") pod \"route-controller-manager-5b6994579d-4mj7f\" (UID: \"f713c391-bba5-4d61-a342-bebc6757fd97\") " pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.472124 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1-proxy-ca-bundles\") pod \"controller-manager-7f5d9b8d98-pwz4t\" (UID: \"b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1\") " pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.472216 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1-client-ca\") pod \"controller-manager-7f5d9b8d98-pwz4t\" (UID: \"b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1\") " pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.472315 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f713c391-bba5-4d61-a342-bebc6757fd97-serving-cert\") pod \"route-controller-manager-5b6994579d-4mj7f\" (UID: \"f713c391-bba5-4d61-a342-bebc6757fd97\") " pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.472402 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1-config\") pod \"controller-manager-7f5d9b8d98-pwz4t\" (UID: \"b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1\") " pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.472480 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1-serving-cert\") pod \"controller-manager-7f5d9b8d98-pwz4t\" (UID: \"b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1\") " pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.472581 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf8sq\" (UniqueName: \"kubernetes.io/projected/b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1-kube-api-access-lf8sq\") pod \"controller-manager-7f5d9b8d98-pwz4t\" (UID: \"b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1\") " pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.472656 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f713c391-bba5-4d61-a342-bebc6757fd97-client-ca\") pod \"route-controller-manager-5b6994579d-4mj7f\" (UID: \"f713c391-bba5-4d61-a342-bebc6757fd97\") " pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.473220 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1-client-ca\") pod \"controller-manager-7f5d9b8d98-pwz4t\" (UID: \"b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1\") " pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.473300 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f713c391-bba5-4d61-a342-bebc6757fd97-config\") pod \"route-controller-manager-5b6994579d-4mj7f\" (UID: \"f713c391-bba5-4d61-a342-bebc6757fd97\") " pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.473731 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1-proxy-ca-bundles\") pod \"controller-manager-7f5d9b8d98-pwz4t\" (UID: \"b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1\") " pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.473966 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1-config\") pod \"controller-manager-7f5d9b8d98-pwz4t\" (UID: \"b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1\") " pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.478166 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f713c391-bba5-4d61-a342-bebc6757fd97-serving-cert\") pod \"route-controller-manager-5b6994579d-4mj7f\" (UID: \"f713c391-bba5-4d61-a342-bebc6757fd97\") " pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.478277 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1-serving-cert\") pod \"controller-manager-7f5d9b8d98-pwz4t\" (UID: \"b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1\") " pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.488446 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lsrq\" (UniqueName: \"kubernetes.io/projected/f713c391-bba5-4d61-a342-bebc6757fd97-kube-api-access-9lsrq\") pod \"route-controller-manager-5b6994579d-4mj7f\" (UID: \"f713c391-bba5-4d61-a342-bebc6757fd97\") " pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.490311 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf8sq\" (UniqueName: \"kubernetes.io/projected/b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1-kube-api-access-lf8sq\") pod \"controller-manager-7f5d9b8d98-pwz4t\" (UID: \"b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1\") " pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.520610 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.526295 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.540431 4828 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.557962 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.564712 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.615140 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.680612 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.687400 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.694752 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.697562 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.745938 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.750893 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t"] Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.761426 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.784059 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.806069 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f"] Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.808003 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 10 18:59:53 crc kubenswrapper[4828]: I1210 18:59:53.820021 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.213162 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.219125 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.219620 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.225019 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.388738 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.462414 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.523027 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.562711 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.628017 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" event={"ID":"f713c391-bba5-4d61-a342-bebc6757fd97","Type":"ContainerStarted","Data":"0e302101d5dd7f31cb5f8c75a31ee67b136602def656b01ce076885b26ca0c4b"} Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.628069 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" event={"ID":"f713c391-bba5-4d61-a342-bebc6757fd97","Type":"ContainerStarted","Data":"cb79d30edc76d353303c3477749644f7e174ab8f778205845630f39bc05905e3"} Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.628361 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.629945 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" event={"ID":"b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1","Type":"ContainerStarted","Data":"5082b4bc09a24f1b97ea390b47f3f789e75dc8da67a985ac4525651f53392b22"} Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.629982 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" event={"ID":"b0a9abc6-45d5-47ce-9a0f-f361ed62bdb1","Type":"ContainerStarted","Data":"ff1c7eaf3767804b8a5e1289a2dd97734865e3be9687309f1c5cd6c5da9527e6"} Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.630212 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.634142 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.635223 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.645788 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5b6994579d-4mj7f" podStartSLOduration=53.645764364 podStartE2EDuration="53.645764364s" podCreationTimestamp="2025-12-10 18:59:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:59:54.643569803 +0000 UTC m=+275.154180808" watchObservedRunningTime="2025-12-10 18:59:54.645764364 +0000 UTC m=+275.156375369" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.664898 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.672500 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7f5d9b8d98-pwz4t" podStartSLOduration=53.672478011 podStartE2EDuration="53.672478011s" podCreationTimestamp="2025-12-10 18:59:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:59:54.665470578 +0000 UTC m=+275.176081583" watchObservedRunningTime="2025-12-10 18:59:54.672478011 +0000 UTC m=+275.183089016" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.683567 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.752072 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.779044 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.781676 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.902643 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.961687 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 10 18:59:54 crc kubenswrapper[4828]: I1210 18:59:54.979482 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.028070 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.385761 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.390953 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.435683 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.450208 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.458276 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.474075 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.490424 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.536874 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.547516 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.571193 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.583357 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.592115 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.669249 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.677378 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.770567 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.790379 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.842352 4828 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.842783 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://1c65abebabfa38770059d8a3ee3bad6c5cafea58415c02a7eaf3efcbbdbb797c" gracePeriod=5 Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.923751 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 10 18:59:55 crc kubenswrapper[4828]: I1210 18:59:55.925632 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 10 18:59:56 crc kubenswrapper[4828]: I1210 18:59:56.213721 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 10 18:59:56 crc kubenswrapper[4828]: I1210 18:59:56.240619 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 10 18:59:56 crc kubenswrapper[4828]: I1210 18:59:56.263692 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 10 18:59:56 crc kubenswrapper[4828]: I1210 18:59:56.291285 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 10 18:59:56 crc kubenswrapper[4828]: I1210 18:59:56.373461 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 10 18:59:56 crc kubenswrapper[4828]: I1210 18:59:56.375875 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 10 18:59:56 crc kubenswrapper[4828]: I1210 18:59:56.450956 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 10 18:59:56 crc kubenswrapper[4828]: I1210 18:59:56.523094 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 10 18:59:56 crc kubenswrapper[4828]: I1210 18:59:56.565763 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 10 18:59:56 crc kubenswrapper[4828]: I1210 18:59:56.588778 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 10 18:59:56 crc kubenswrapper[4828]: I1210 18:59:56.773135 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 10 18:59:56 crc kubenswrapper[4828]: I1210 18:59:56.832626 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 10 18:59:56 crc kubenswrapper[4828]: I1210 18:59:56.881769 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 10 18:59:56 crc kubenswrapper[4828]: I1210 18:59:56.925584 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 10 18:59:56 crc kubenswrapper[4828]: I1210 18:59:56.926389 4828 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 10 18:59:56 crc kubenswrapper[4828]: I1210 18:59:56.970000 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 10 18:59:57 crc kubenswrapper[4828]: I1210 18:59:57.006918 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 10 18:59:57 crc kubenswrapper[4828]: I1210 18:59:57.019426 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 10 18:59:57 crc kubenswrapper[4828]: I1210 18:59:57.183440 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 10 18:59:57 crc kubenswrapper[4828]: I1210 18:59:57.240894 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 10 18:59:57 crc kubenswrapper[4828]: I1210 18:59:57.284662 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 10 18:59:57 crc kubenswrapper[4828]: I1210 18:59:57.295393 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 10 18:59:57 crc kubenswrapper[4828]: I1210 18:59:57.326669 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 10 18:59:57 crc kubenswrapper[4828]: I1210 18:59:57.413517 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 10 18:59:57 crc kubenswrapper[4828]: I1210 18:59:57.499216 4828 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 10 18:59:57 crc kubenswrapper[4828]: I1210 18:59:57.511249 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 10 18:59:57 crc kubenswrapper[4828]: I1210 18:59:57.520775 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 10 18:59:57 crc kubenswrapper[4828]: I1210 18:59:57.618881 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 10 18:59:57 crc kubenswrapper[4828]: I1210 18:59:57.623126 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 10 18:59:57 crc kubenswrapper[4828]: I1210 18:59:57.625397 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 10 18:59:57 crc kubenswrapper[4828]: I1210 18:59:57.787225 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 10 18:59:57 crc kubenswrapper[4828]: I1210 18:59:57.801289 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 10 18:59:57 crc kubenswrapper[4828]: I1210 18:59:57.989480 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 10 18:59:58 crc kubenswrapper[4828]: I1210 18:59:58.038228 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 10 18:59:58 crc kubenswrapper[4828]: I1210 18:59:58.052713 4828 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 10 18:59:58 crc kubenswrapper[4828]: I1210 18:59:58.069560 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 10 18:59:58 crc kubenswrapper[4828]: I1210 18:59:58.144395 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 10 18:59:58 crc kubenswrapper[4828]: I1210 18:59:58.237757 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 10 18:59:58 crc kubenswrapper[4828]: I1210 18:59:58.308009 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 10 18:59:58 crc kubenswrapper[4828]: I1210 18:59:58.400109 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 10 18:59:58 crc kubenswrapper[4828]: I1210 18:59:58.402150 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 10 18:59:58 crc kubenswrapper[4828]: I1210 18:59:58.431718 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 10 18:59:58 crc kubenswrapper[4828]: I1210 18:59:58.788475 4828 scope.go:117] "RemoveContainer" containerID="109b9cc57062b7758b719816a1707eb10ea6545c9865b1babb0d2b0fb3dc4048" Dec 10 18:59:59 crc kubenswrapper[4828]: I1210 18:59:59.012759 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 10 18:59:59 crc kubenswrapper[4828]: I1210 18:59:59.340886 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 10 18:59:59 crc kubenswrapper[4828]: I1210 18:59:59.549868 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 10 18:59:59 crc kubenswrapper[4828]: I1210 18:59:59.660731 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-7d9549f6c-gj4pr_8ab5a672-80db-49a0-9b83-db5b7a5b4e65/oauth-openshift/1.log" Dec 10 18:59:59 crc kubenswrapper[4828]: I1210 18:59:59.660896 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" event={"ID":"8ab5a672-80db-49a0-9b83-db5b7a5b4e65","Type":"ContainerStarted","Data":"fe6198015d1d5db45557179f22c451000274df37afada6daeac00bcb314b124c"} Dec 10 18:59:59 crc kubenswrapper[4828]: I1210 18:59:59.661327 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:59 crc kubenswrapper[4828]: I1210 18:59:59.674764 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" Dec 10 18:59:59 crc kubenswrapper[4828]: I1210 18:59:59.687144 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7d9549f6c-gj4pr" podStartSLOduration=91.68712265 podStartE2EDuration="1m31.68712265s" podCreationTimestamp="2025-12-10 18:58:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 18:59:38.5523216 +0000 UTC m=+259.062932615" watchObservedRunningTime="2025-12-10 18:59:59.68712265 +0000 UTC m=+280.197733655" Dec 10 18:59:59 crc kubenswrapper[4828]: I1210 18:59:59.957738 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.174501 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd"] Dec 10 19:00:00 crc kubenswrapper[4828]: E1210 19:00:00.174859 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.174883 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.175070 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.175712 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.177837 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.178953 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.188432 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd"] Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.258720 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pp8h\" (UniqueName: \"kubernetes.io/projected/22097d8d-0d50-4948-984b-4ea71356df75-kube-api-access-6pp8h\") pod \"collect-profiles-29423220-8r8rd\" (UID: \"22097d8d-0d50-4948-984b-4ea71356df75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.258876 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/22097d8d-0d50-4948-984b-4ea71356df75-secret-volume\") pod \"collect-profiles-29423220-8r8rd\" (UID: \"22097d8d-0d50-4948-984b-4ea71356df75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.258923 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22097d8d-0d50-4948-984b-4ea71356df75-config-volume\") pod \"collect-profiles-29423220-8r8rd\" (UID: \"22097d8d-0d50-4948-984b-4ea71356df75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.360528 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22097d8d-0d50-4948-984b-4ea71356df75-config-volume\") pod \"collect-profiles-29423220-8r8rd\" (UID: \"22097d8d-0d50-4948-984b-4ea71356df75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.360625 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pp8h\" (UniqueName: \"kubernetes.io/projected/22097d8d-0d50-4948-984b-4ea71356df75-kube-api-access-6pp8h\") pod \"collect-profiles-29423220-8r8rd\" (UID: \"22097d8d-0d50-4948-984b-4ea71356df75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.360716 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/22097d8d-0d50-4948-984b-4ea71356df75-secret-volume\") pod \"collect-profiles-29423220-8r8rd\" (UID: \"22097d8d-0d50-4948-984b-4ea71356df75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.364112 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22097d8d-0d50-4948-984b-4ea71356df75-config-volume\") pod \"collect-profiles-29423220-8r8rd\" (UID: \"22097d8d-0d50-4948-984b-4ea71356df75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.370381 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/22097d8d-0d50-4948-984b-4ea71356df75-secret-volume\") pod \"collect-profiles-29423220-8r8rd\" (UID: \"22097d8d-0d50-4948-984b-4ea71356df75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.379764 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pp8h\" (UniqueName: \"kubernetes.io/projected/22097d8d-0d50-4948-984b-4ea71356df75-kube-api-access-6pp8h\") pod \"collect-profiles-29423220-8r8rd\" (UID: \"22097d8d-0d50-4948-984b-4ea71356df75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.487093 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.490909 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd" Dec 10 19:00:00 crc kubenswrapper[4828]: I1210 19:00:00.884767 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd"] Dec 10 19:00:00 crc kubenswrapper[4828]: W1210 19:00:00.940547 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22097d8d_0d50_4948_984b_4ea71356df75.slice/crio-f6e6088db2786d5a34a98afab6ed7a9bc582b3eba15944102a570d6e9538b26b WatchSource:0}: Error finding container f6e6088db2786d5a34a98afab6ed7a9bc582b3eba15944102a570d6e9538b26b: Status 404 returned error can't find the container with id f6e6088db2786d5a34a98afab6ed7a9bc582b3eba15944102a570d6e9538b26b Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.674730 4828 generic.go:334] "Generic (PLEG): container finished" podID="22097d8d-0d50-4948-984b-4ea71356df75" containerID="b0db0480ae039902518f264b8731bfb8c9a45bb71ef1a6b105377e32771ef083" exitCode=0 Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.674812 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd" event={"ID":"22097d8d-0d50-4948-984b-4ea71356df75","Type":"ContainerDied","Data":"b0db0480ae039902518f264b8731bfb8c9a45bb71ef1a6b105377e32771ef083"} Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.675123 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd" event={"ID":"22097d8d-0d50-4948-984b-4ea71356df75","Type":"ContainerStarted","Data":"f6e6088db2786d5a34a98afab6ed7a9bc582b3eba15944102a570d6e9538b26b"} Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.677473 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.677517 4828 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="1c65abebabfa38770059d8a3ee3bad6c5cafea58415c02a7eaf3efcbbdbb797c" exitCode=137 Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.677723 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="def7bc69e74757d5f71a31b5351910286f1121773995bb23241de3a760e742bd" Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.684450 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.684506 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.777094 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.777170 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.777197 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.777255 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.777261 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.777283 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.777290 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.777324 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.777419 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.777693 4828 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.777711 4828 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.777719 4828 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.777730 4828 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.785439 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.795347 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 10 19:00:01 crc kubenswrapper[4828]: I1210 19:00:01.878641 4828 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:02 crc kubenswrapper[4828]: I1210 19:00:02.682856 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 10 19:00:03 crc kubenswrapper[4828]: I1210 19:00:03.039282 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd" Dec 10 19:00:03 crc kubenswrapper[4828]: I1210 19:00:03.094196 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22097d8d-0d50-4948-984b-4ea71356df75-config-volume\") pod \"22097d8d-0d50-4948-984b-4ea71356df75\" (UID: \"22097d8d-0d50-4948-984b-4ea71356df75\") " Dec 10 19:00:03 crc kubenswrapper[4828]: I1210 19:00:03.094332 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/22097d8d-0d50-4948-984b-4ea71356df75-secret-volume\") pod \"22097d8d-0d50-4948-984b-4ea71356df75\" (UID: \"22097d8d-0d50-4948-984b-4ea71356df75\") " Dec 10 19:00:03 crc kubenswrapper[4828]: I1210 19:00:03.094441 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pp8h\" (UniqueName: \"kubernetes.io/projected/22097d8d-0d50-4948-984b-4ea71356df75-kube-api-access-6pp8h\") pod \"22097d8d-0d50-4948-984b-4ea71356df75\" (UID: \"22097d8d-0d50-4948-984b-4ea71356df75\") " Dec 10 19:00:03 crc kubenswrapper[4828]: I1210 19:00:03.095032 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22097d8d-0d50-4948-984b-4ea71356df75-config-volume" (OuterVolumeSpecName: "config-volume") pod "22097d8d-0d50-4948-984b-4ea71356df75" (UID: "22097d8d-0d50-4948-984b-4ea71356df75"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:00:03 crc kubenswrapper[4828]: I1210 19:00:03.098492 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22097d8d-0d50-4948-984b-4ea71356df75-kube-api-access-6pp8h" (OuterVolumeSpecName: "kube-api-access-6pp8h") pod "22097d8d-0d50-4948-984b-4ea71356df75" (UID: "22097d8d-0d50-4948-984b-4ea71356df75"). InnerVolumeSpecName "kube-api-access-6pp8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:00:03 crc kubenswrapper[4828]: I1210 19:00:03.100036 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22097d8d-0d50-4948-984b-4ea71356df75-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "22097d8d-0d50-4948-984b-4ea71356df75" (UID: "22097d8d-0d50-4948-984b-4ea71356df75"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:00:03 crc kubenswrapper[4828]: I1210 19:00:03.196262 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pp8h\" (UniqueName: \"kubernetes.io/projected/22097d8d-0d50-4948-984b-4ea71356df75-kube-api-access-6pp8h\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:03 crc kubenswrapper[4828]: I1210 19:00:03.196311 4828 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/22097d8d-0d50-4948-984b-4ea71356df75-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:03 crc kubenswrapper[4828]: I1210 19:00:03.196333 4828 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/22097d8d-0d50-4948-984b-4ea71356df75-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:03 crc kubenswrapper[4828]: I1210 19:00:03.691017 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd" event={"ID":"22097d8d-0d50-4948-984b-4ea71356df75","Type":"ContainerDied","Data":"f6e6088db2786d5a34a98afab6ed7a9bc582b3eba15944102a570d6e9538b26b"} Dec 10 19:00:03 crc kubenswrapper[4828]: I1210 19:00:03.691081 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6e6088db2786d5a34a98afab6ed7a9bc582b3eba15944102a570d6e9538b26b" Dec 10 19:00:03 crc kubenswrapper[4828]: I1210 19:00:03.691172 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd" Dec 10 19:00:16 crc kubenswrapper[4828]: I1210 19:00:16.764253 4828 generic.go:334] "Generic (PLEG): container finished" podID="c5bd6649-abb0-4898-a73f-efceb62d2cb9" containerID="2d897701418eef8ce32adfb032b1a0b184b7fe03ec55abcb747335b88b3ebb7d" exitCode=0 Dec 10 19:00:16 crc kubenswrapper[4828]: I1210 19:00:16.764349 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" event={"ID":"c5bd6649-abb0-4898-a73f-efceb62d2cb9","Type":"ContainerDied","Data":"2d897701418eef8ce32adfb032b1a0b184b7fe03ec55abcb747335b88b3ebb7d"} Dec 10 19:00:16 crc kubenswrapper[4828]: I1210 19:00:16.765243 4828 scope.go:117] "RemoveContainer" containerID="2d897701418eef8ce32adfb032b1a0b184b7fe03ec55abcb747335b88b3ebb7d" Dec 10 19:00:17 crc kubenswrapper[4828]: I1210 19:00:17.771644 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" event={"ID":"c5bd6649-abb0-4898-a73f-efceb62d2cb9","Type":"ContainerStarted","Data":"268da9959d149cddc863357066862e37a4f201fdd5fda2fc8175ac3266496ac9"} Dec 10 19:00:17 crc kubenswrapper[4828]: I1210 19:00:17.772424 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" Dec 10 19:00:17 crc kubenswrapper[4828]: I1210 19:00:17.773761 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" Dec 10 19:00:19 crc kubenswrapper[4828]: I1210 19:00:19.651108 4828 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 10 19:00:29 crc kubenswrapper[4828]: I1210 19:00:29.366464 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.400224 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2vvgb"] Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.401774 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2vvgb" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" containerName="registry-server" containerID="cri-o://b64b11b0c4a38dfdbb4d958370a441a18ed178d13810ae372fca0e3827a6cc80" gracePeriod=2 Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.599824 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ll5gw"] Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.600106 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ll5gw" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" containerName="registry-server" containerID="cri-o://2a1381f205225da9f5c6e9cc8aa98f6addf271ef62ce9e5dccf0ab9021234baf" gracePeriod=2 Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.787416 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2vvgb" Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.840209 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6z2dm\" (UniqueName: \"kubernetes.io/projected/43cd3400-d5e9-4279-9a85-051625ca4607-kube-api-access-6z2dm\") pod \"43cd3400-d5e9-4279-9a85-051625ca4607\" (UID: \"43cd3400-d5e9-4279-9a85-051625ca4607\") " Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.840250 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43cd3400-d5e9-4279-9a85-051625ca4607-utilities\") pod \"43cd3400-d5e9-4279-9a85-051625ca4607\" (UID: \"43cd3400-d5e9-4279-9a85-051625ca4607\") " Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.840316 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43cd3400-d5e9-4279-9a85-051625ca4607-catalog-content\") pod \"43cd3400-d5e9-4279-9a85-051625ca4607\" (UID: \"43cd3400-d5e9-4279-9a85-051625ca4607\") " Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.841520 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43cd3400-d5e9-4279-9a85-051625ca4607-utilities" (OuterVolumeSpecName: "utilities") pod "43cd3400-d5e9-4279-9a85-051625ca4607" (UID: "43cd3400-d5e9-4279-9a85-051625ca4607"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.861062 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43cd3400-d5e9-4279-9a85-051625ca4607-kube-api-access-6z2dm" (OuterVolumeSpecName: "kube-api-access-6z2dm") pod "43cd3400-d5e9-4279-9a85-051625ca4607" (UID: "43cd3400-d5e9-4279-9a85-051625ca4607"). InnerVolumeSpecName "kube-api-access-6z2dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.892932 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43cd3400-d5e9-4279-9a85-051625ca4607-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "43cd3400-d5e9-4279-9a85-051625ca4607" (UID: "43cd3400-d5e9-4279-9a85-051625ca4607"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.941926 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43cd3400-d5e9-4279-9a85-051625ca4607-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.941982 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6z2dm\" (UniqueName: \"kubernetes.io/projected/43cd3400-d5e9-4279-9a85-051625ca4607-kube-api-access-6z2dm\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.942000 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43cd3400-d5e9-4279-9a85-051625ca4607-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.947919 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ll5gw" Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.975614 4828 generic.go:334] "Generic (PLEG): container finished" podID="43cd3400-d5e9-4279-9a85-051625ca4607" containerID="b64b11b0c4a38dfdbb4d958370a441a18ed178d13810ae372fca0e3827a6cc80" exitCode=0 Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.975682 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2vvgb" event={"ID":"43cd3400-d5e9-4279-9a85-051625ca4607","Type":"ContainerDied","Data":"b64b11b0c4a38dfdbb4d958370a441a18ed178d13810ae372fca0e3827a6cc80"} Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.975710 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2vvgb" event={"ID":"43cd3400-d5e9-4279-9a85-051625ca4607","Type":"ContainerDied","Data":"496dfddc50d3cc05bbad0965d627b7e72f2799b2fb422c04d246b4ea745cd1b5"} Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.975726 4828 scope.go:117] "RemoveContainer" containerID="b64b11b0c4a38dfdbb4d958370a441a18ed178d13810ae372fca0e3827a6cc80" Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.975849 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2vvgb" Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.979845 4828 generic.go:334] "Generic (PLEG): container finished" podID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" containerID="2a1381f205225da9f5c6e9cc8aa98f6addf271ef62ce9e5dccf0ab9021234baf" exitCode=0 Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.979889 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ll5gw" event={"ID":"44b88622-40eb-4bcf-81a5-dab1f8b4c63b","Type":"ContainerDied","Data":"2a1381f205225da9f5c6e9cc8aa98f6addf271ef62ce9e5dccf0ab9021234baf"} Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.979919 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ll5gw" event={"ID":"44b88622-40eb-4bcf-81a5-dab1f8b4c63b","Type":"ContainerDied","Data":"253088b4f1912303196e311ca62da10c9615c2eb2b30772f92278d42a79e2264"} Dec 10 19:00:53 crc kubenswrapper[4828]: I1210 19:00:53.979944 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ll5gw" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.001474 4828 scope.go:117] "RemoveContainer" containerID="566f38d4316412e795518f193b1edf77c2317782cc06cd4dea00d0fa0b5f882f" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.005944 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2vvgb"] Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.014562 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2vvgb"] Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.023054 4828 scope.go:117] "RemoveContainer" containerID="442bd248c218c33c47b0aec9abca9ea73132a233d7290ac75c33dd4c60ee9ced" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.035869 4828 scope.go:117] "RemoveContainer" containerID="b64b11b0c4a38dfdbb4d958370a441a18ed178d13810ae372fca0e3827a6cc80" Dec 10 19:00:54 crc kubenswrapper[4828]: E1210 19:00:54.036409 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b64b11b0c4a38dfdbb4d958370a441a18ed178d13810ae372fca0e3827a6cc80\": container with ID starting with b64b11b0c4a38dfdbb4d958370a441a18ed178d13810ae372fca0e3827a6cc80 not found: ID does not exist" containerID="b64b11b0c4a38dfdbb4d958370a441a18ed178d13810ae372fca0e3827a6cc80" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.036455 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b64b11b0c4a38dfdbb4d958370a441a18ed178d13810ae372fca0e3827a6cc80"} err="failed to get container status \"b64b11b0c4a38dfdbb4d958370a441a18ed178d13810ae372fca0e3827a6cc80\": rpc error: code = NotFound desc = could not find container \"b64b11b0c4a38dfdbb4d958370a441a18ed178d13810ae372fca0e3827a6cc80\": container with ID starting with b64b11b0c4a38dfdbb4d958370a441a18ed178d13810ae372fca0e3827a6cc80 not found: ID does not exist" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.036488 4828 scope.go:117] "RemoveContainer" containerID="566f38d4316412e795518f193b1edf77c2317782cc06cd4dea00d0fa0b5f882f" Dec 10 19:00:54 crc kubenswrapper[4828]: E1210 19:00:54.036762 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"566f38d4316412e795518f193b1edf77c2317782cc06cd4dea00d0fa0b5f882f\": container with ID starting with 566f38d4316412e795518f193b1edf77c2317782cc06cd4dea00d0fa0b5f882f not found: ID does not exist" containerID="566f38d4316412e795518f193b1edf77c2317782cc06cd4dea00d0fa0b5f882f" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.036790 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"566f38d4316412e795518f193b1edf77c2317782cc06cd4dea00d0fa0b5f882f"} err="failed to get container status \"566f38d4316412e795518f193b1edf77c2317782cc06cd4dea00d0fa0b5f882f\": rpc error: code = NotFound desc = could not find container \"566f38d4316412e795518f193b1edf77c2317782cc06cd4dea00d0fa0b5f882f\": container with ID starting with 566f38d4316412e795518f193b1edf77c2317782cc06cd4dea00d0fa0b5f882f not found: ID does not exist" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.036831 4828 scope.go:117] "RemoveContainer" containerID="442bd248c218c33c47b0aec9abca9ea73132a233d7290ac75c33dd4c60ee9ced" Dec 10 19:00:54 crc kubenswrapper[4828]: E1210 19:00:54.037050 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"442bd248c218c33c47b0aec9abca9ea73132a233d7290ac75c33dd4c60ee9ced\": container with ID starting with 442bd248c218c33c47b0aec9abca9ea73132a233d7290ac75c33dd4c60ee9ced not found: ID does not exist" containerID="442bd248c218c33c47b0aec9abca9ea73132a233d7290ac75c33dd4c60ee9ced" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.037104 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"442bd248c218c33c47b0aec9abca9ea73132a233d7290ac75c33dd4c60ee9ced"} err="failed to get container status \"442bd248c218c33c47b0aec9abca9ea73132a233d7290ac75c33dd4c60ee9ced\": rpc error: code = NotFound desc = could not find container \"442bd248c218c33c47b0aec9abca9ea73132a233d7290ac75c33dd4c60ee9ced\": container with ID starting with 442bd248c218c33c47b0aec9abca9ea73132a233d7290ac75c33dd4c60ee9ced not found: ID does not exist" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.037123 4828 scope.go:117] "RemoveContainer" containerID="2a1381f205225da9f5c6e9cc8aa98f6addf271ef62ce9e5dccf0ab9021234baf" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.043161 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-catalog-content\") pod \"44b88622-40eb-4bcf-81a5-dab1f8b4c63b\" (UID: \"44b88622-40eb-4bcf-81a5-dab1f8b4c63b\") " Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.043265 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-utilities\") pod \"44b88622-40eb-4bcf-81a5-dab1f8b4c63b\" (UID: \"44b88622-40eb-4bcf-81a5-dab1f8b4c63b\") " Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.043343 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7gp8\" (UniqueName: \"kubernetes.io/projected/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-kube-api-access-w7gp8\") pod \"44b88622-40eb-4bcf-81a5-dab1f8b4c63b\" (UID: \"44b88622-40eb-4bcf-81a5-dab1f8b4c63b\") " Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.044251 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-utilities" (OuterVolumeSpecName: "utilities") pod "44b88622-40eb-4bcf-81a5-dab1f8b4c63b" (UID: "44b88622-40eb-4bcf-81a5-dab1f8b4c63b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.046059 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-kube-api-access-w7gp8" (OuterVolumeSpecName: "kube-api-access-w7gp8") pod "44b88622-40eb-4bcf-81a5-dab1f8b4c63b" (UID: "44b88622-40eb-4bcf-81a5-dab1f8b4c63b"). InnerVolumeSpecName "kube-api-access-w7gp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.050158 4828 scope.go:117] "RemoveContainer" containerID="a36f4ae4f7d89b2d0700e44d25d272f7e302cc3d88762fe33c8e233f20c584f2" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.073777 4828 scope.go:117] "RemoveContainer" containerID="d7b83a7de5179e88d2fd506f49860f038da6427203204ff64cdf56a978dc792e" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.089077 4828 scope.go:117] "RemoveContainer" containerID="2a1381f205225da9f5c6e9cc8aa98f6addf271ef62ce9e5dccf0ab9021234baf" Dec 10 19:00:54 crc kubenswrapper[4828]: E1210 19:00:54.089498 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a1381f205225da9f5c6e9cc8aa98f6addf271ef62ce9e5dccf0ab9021234baf\": container with ID starting with 2a1381f205225da9f5c6e9cc8aa98f6addf271ef62ce9e5dccf0ab9021234baf not found: ID does not exist" containerID="2a1381f205225da9f5c6e9cc8aa98f6addf271ef62ce9e5dccf0ab9021234baf" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.089542 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a1381f205225da9f5c6e9cc8aa98f6addf271ef62ce9e5dccf0ab9021234baf"} err="failed to get container status \"2a1381f205225da9f5c6e9cc8aa98f6addf271ef62ce9e5dccf0ab9021234baf\": rpc error: code = NotFound desc = could not find container \"2a1381f205225da9f5c6e9cc8aa98f6addf271ef62ce9e5dccf0ab9021234baf\": container with ID starting with 2a1381f205225da9f5c6e9cc8aa98f6addf271ef62ce9e5dccf0ab9021234baf not found: ID does not exist" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.089598 4828 scope.go:117] "RemoveContainer" containerID="a36f4ae4f7d89b2d0700e44d25d272f7e302cc3d88762fe33c8e233f20c584f2" Dec 10 19:00:54 crc kubenswrapper[4828]: E1210 19:00:54.089926 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a36f4ae4f7d89b2d0700e44d25d272f7e302cc3d88762fe33c8e233f20c584f2\": container with ID starting with a36f4ae4f7d89b2d0700e44d25d272f7e302cc3d88762fe33c8e233f20c584f2 not found: ID does not exist" containerID="a36f4ae4f7d89b2d0700e44d25d272f7e302cc3d88762fe33c8e233f20c584f2" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.089975 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a36f4ae4f7d89b2d0700e44d25d272f7e302cc3d88762fe33c8e233f20c584f2"} err="failed to get container status \"a36f4ae4f7d89b2d0700e44d25d272f7e302cc3d88762fe33c8e233f20c584f2\": rpc error: code = NotFound desc = could not find container \"a36f4ae4f7d89b2d0700e44d25d272f7e302cc3d88762fe33c8e233f20c584f2\": container with ID starting with a36f4ae4f7d89b2d0700e44d25d272f7e302cc3d88762fe33c8e233f20c584f2 not found: ID does not exist" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.090012 4828 scope.go:117] "RemoveContainer" containerID="d7b83a7de5179e88d2fd506f49860f038da6427203204ff64cdf56a978dc792e" Dec 10 19:00:54 crc kubenswrapper[4828]: E1210 19:00:54.090335 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7b83a7de5179e88d2fd506f49860f038da6427203204ff64cdf56a978dc792e\": container with ID starting with d7b83a7de5179e88d2fd506f49860f038da6427203204ff64cdf56a978dc792e not found: ID does not exist" containerID="d7b83a7de5179e88d2fd506f49860f038da6427203204ff64cdf56a978dc792e" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.090447 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7b83a7de5179e88d2fd506f49860f038da6427203204ff64cdf56a978dc792e"} err="failed to get container status \"d7b83a7de5179e88d2fd506f49860f038da6427203204ff64cdf56a978dc792e\": rpc error: code = NotFound desc = could not find container \"d7b83a7de5179e88d2fd506f49860f038da6427203204ff64cdf56a978dc792e\": container with ID starting with d7b83a7de5179e88d2fd506f49860f038da6427203204ff64cdf56a978dc792e not found: ID does not exist" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.097206 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44b88622-40eb-4bcf-81a5-dab1f8b4c63b" (UID: "44b88622-40eb-4bcf-81a5-dab1f8b4c63b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.144211 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.144244 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7gp8\" (UniqueName: \"kubernetes.io/projected/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-kube-api-access-w7gp8\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.144259 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b88622-40eb-4bcf-81a5-dab1f8b4c63b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.308388 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ll5gw"] Dec 10 19:00:54 crc kubenswrapper[4828]: I1210 19:00:54.313186 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ll5gw"] Dec 10 19:00:55 crc kubenswrapper[4828]: I1210 19:00:55.797006 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" path="/var/lib/kubelet/pods/43cd3400-d5e9-4279-9a85-051625ca4607/volumes" Dec 10 19:00:55 crc kubenswrapper[4828]: I1210 19:00:55.798987 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" path="/var/lib/kubelet/pods/44b88622-40eb-4bcf-81a5-dab1f8b4c63b/volumes" Dec 10 19:00:55 crc kubenswrapper[4828]: I1210 19:00:55.803526 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ftr4x"] Dec 10 19:00:55 crc kubenswrapper[4828]: I1210 19:00:55.803757 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ftr4x" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" containerName="registry-server" containerID="cri-o://40cba23895e88153418f64c22b79e8dd857b8c34c57d2e037eedd2053ebe9b65" gracePeriod=2 Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.002718 4828 generic.go:334] "Generic (PLEG): container finished" podID="3ef84f18-e802-4038-ba08-2b4eb948d803" containerID="40cba23895e88153418f64c22b79e8dd857b8c34c57d2e037eedd2053ebe9b65" exitCode=0 Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.002879 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftr4x" event={"ID":"3ef84f18-e802-4038-ba08-2b4eb948d803","Type":"ContainerDied","Data":"40cba23895e88153418f64c22b79e8dd857b8c34c57d2e037eedd2053ebe9b65"} Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.004301 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kfgtf"] Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.004677 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kfgtf" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" containerName="registry-server" containerID="cri-o://e5acfe41e8f64a9ae2b49b3c66e8f1b9b3940b390f2a934974073d6e4afee137" gracePeriod=2 Dec 10 19:00:56 crc kubenswrapper[4828]: E1210 19:00:56.142388 4828 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 40cba23895e88153418f64c22b79e8dd857b8c34c57d2e037eedd2053ebe9b65 is running failed: container process not found" containerID="40cba23895e88153418f64c22b79e8dd857b8c34c57d2e037eedd2053ebe9b65" cmd=["grpc_health_probe","-addr=:50051"] Dec 10 19:00:56 crc kubenswrapper[4828]: E1210 19:00:56.143888 4828 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 40cba23895e88153418f64c22b79e8dd857b8c34c57d2e037eedd2053ebe9b65 is running failed: container process not found" containerID="40cba23895e88153418f64c22b79e8dd857b8c34c57d2e037eedd2053ebe9b65" cmd=["grpc_health_probe","-addr=:50051"] Dec 10 19:00:56 crc kubenswrapper[4828]: E1210 19:00:56.144290 4828 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 40cba23895e88153418f64c22b79e8dd857b8c34c57d2e037eedd2053ebe9b65 is running failed: container process not found" containerID="40cba23895e88153418f64c22b79e8dd857b8c34c57d2e037eedd2053ebe9b65" cmd=["grpc_health_probe","-addr=:50051"] Dec 10 19:00:56 crc kubenswrapper[4828]: E1210 19:00:56.144343 4828 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 40cba23895e88153418f64c22b79e8dd857b8c34c57d2e037eedd2053ebe9b65 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-ftr4x" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" containerName="registry-server" Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.197631 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ftr4x" Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.270067 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ef84f18-e802-4038-ba08-2b4eb948d803-utilities\") pod \"3ef84f18-e802-4038-ba08-2b4eb948d803\" (UID: \"3ef84f18-e802-4038-ba08-2b4eb948d803\") " Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.270139 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg6cp\" (UniqueName: \"kubernetes.io/projected/3ef84f18-e802-4038-ba08-2b4eb948d803-kube-api-access-wg6cp\") pod \"3ef84f18-e802-4038-ba08-2b4eb948d803\" (UID: \"3ef84f18-e802-4038-ba08-2b4eb948d803\") " Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.270192 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ef84f18-e802-4038-ba08-2b4eb948d803-catalog-content\") pod \"3ef84f18-e802-4038-ba08-2b4eb948d803\" (UID: \"3ef84f18-e802-4038-ba08-2b4eb948d803\") " Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.271170 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ef84f18-e802-4038-ba08-2b4eb948d803-utilities" (OuterVolumeSpecName: "utilities") pod "3ef84f18-e802-4038-ba08-2b4eb948d803" (UID: "3ef84f18-e802-4038-ba08-2b4eb948d803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.275004 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ef84f18-e802-4038-ba08-2b4eb948d803-kube-api-access-wg6cp" (OuterVolumeSpecName: "kube-api-access-wg6cp") pod "3ef84f18-e802-4038-ba08-2b4eb948d803" (UID: "3ef84f18-e802-4038-ba08-2b4eb948d803"). InnerVolumeSpecName "kube-api-access-wg6cp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.293465 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ef84f18-e802-4038-ba08-2b4eb948d803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3ef84f18-e802-4038-ba08-2b4eb948d803" (UID: "3ef84f18-e802-4038-ba08-2b4eb948d803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.356983 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kfgtf" Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.374224 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ef84f18-e802-4038-ba08-2b4eb948d803-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.374266 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg6cp\" (UniqueName: \"kubernetes.io/projected/3ef84f18-e802-4038-ba08-2b4eb948d803-kube-api-access-wg6cp\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.374279 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ef84f18-e802-4038-ba08-2b4eb948d803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.474899 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hg6qz\" (UniqueName: \"kubernetes.io/projected/e466f0bb-f597-46e1-b27a-a361322a6dcd-kube-api-access-hg6qz\") pod \"e466f0bb-f597-46e1-b27a-a361322a6dcd\" (UID: \"e466f0bb-f597-46e1-b27a-a361322a6dcd\") " Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.474997 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e466f0bb-f597-46e1-b27a-a361322a6dcd-utilities\") pod \"e466f0bb-f597-46e1-b27a-a361322a6dcd\" (UID: \"e466f0bb-f597-46e1-b27a-a361322a6dcd\") " Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.475032 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e466f0bb-f597-46e1-b27a-a361322a6dcd-catalog-content\") pod \"e466f0bb-f597-46e1-b27a-a361322a6dcd\" (UID: \"e466f0bb-f597-46e1-b27a-a361322a6dcd\") " Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.475926 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e466f0bb-f597-46e1-b27a-a361322a6dcd-utilities" (OuterVolumeSpecName: "utilities") pod "e466f0bb-f597-46e1-b27a-a361322a6dcd" (UID: "e466f0bb-f597-46e1-b27a-a361322a6dcd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.480157 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e466f0bb-f597-46e1-b27a-a361322a6dcd-kube-api-access-hg6qz" (OuterVolumeSpecName: "kube-api-access-hg6qz") pod "e466f0bb-f597-46e1-b27a-a361322a6dcd" (UID: "e466f0bb-f597-46e1-b27a-a361322a6dcd"). InnerVolumeSpecName "kube-api-access-hg6qz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.576035 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hg6qz\" (UniqueName: \"kubernetes.io/projected/e466f0bb-f597-46e1-b27a-a361322a6dcd-kube-api-access-hg6qz\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.576064 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e466f0bb-f597-46e1-b27a-a361322a6dcd-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.605295 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e466f0bb-f597-46e1-b27a-a361322a6dcd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e466f0bb-f597-46e1-b27a-a361322a6dcd" (UID: "e466f0bb-f597-46e1-b27a-a361322a6dcd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:00:56 crc kubenswrapper[4828]: I1210 19:00:56.676757 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e466f0bb-f597-46e1-b27a-a361322a6dcd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.009470 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ftr4x" event={"ID":"3ef84f18-e802-4038-ba08-2b4eb948d803","Type":"ContainerDied","Data":"c60c7f9b2d1ddd727493825a7f9e6a3f74d3ddf259f324b4f791740a2807dffe"} Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.009537 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ftr4x" Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.010583 4828 scope.go:117] "RemoveContainer" containerID="40cba23895e88153418f64c22b79e8dd857b8c34c57d2e037eedd2053ebe9b65" Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.012285 4828 generic.go:334] "Generic (PLEG): container finished" podID="e466f0bb-f597-46e1-b27a-a361322a6dcd" containerID="e5acfe41e8f64a9ae2b49b3c66e8f1b9b3940b390f2a934974073d6e4afee137" exitCode=0 Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.012340 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kfgtf" Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.012338 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfgtf" event={"ID":"e466f0bb-f597-46e1-b27a-a361322a6dcd","Type":"ContainerDied","Data":"e5acfe41e8f64a9ae2b49b3c66e8f1b9b3940b390f2a934974073d6e4afee137"} Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.013216 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kfgtf" event={"ID":"e466f0bb-f597-46e1-b27a-a361322a6dcd","Type":"ContainerDied","Data":"4750b46b74dfac1ddaa41feb82e2e92bd3c7c970e69a1ff9cbda62d7f28b35bb"} Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.032774 4828 scope.go:117] "RemoveContainer" containerID="a8862133845c223cf7641c8e31cbade95eddb6d7fda7b09a42edf8bf3e3c86e5" Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.040357 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kfgtf"] Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.043190 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kfgtf"] Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.052227 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ftr4x"] Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.057209 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ftr4x"] Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.062024 4828 scope.go:117] "RemoveContainer" containerID="c7959dcffd3e6f35062fb5118ec6fc089830b0634d9bdbc808f808ce9aa7d801" Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.082470 4828 scope.go:117] "RemoveContainer" containerID="e5acfe41e8f64a9ae2b49b3c66e8f1b9b3940b390f2a934974073d6e4afee137" Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.099511 4828 scope.go:117] "RemoveContainer" containerID="f38682501ce035cb7c913628c69c4d62174d66e1903f367119c8168eac7d4095" Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.113571 4828 scope.go:117] "RemoveContainer" containerID="32acb243cd59bea695a829412e491cd934342d0e3457c6e66993db1cc091c887" Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.134138 4828 scope.go:117] "RemoveContainer" containerID="e5acfe41e8f64a9ae2b49b3c66e8f1b9b3940b390f2a934974073d6e4afee137" Dec 10 19:00:57 crc kubenswrapper[4828]: E1210 19:00:57.134651 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5acfe41e8f64a9ae2b49b3c66e8f1b9b3940b390f2a934974073d6e4afee137\": container with ID starting with e5acfe41e8f64a9ae2b49b3c66e8f1b9b3940b390f2a934974073d6e4afee137 not found: ID does not exist" containerID="e5acfe41e8f64a9ae2b49b3c66e8f1b9b3940b390f2a934974073d6e4afee137" Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.134699 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5acfe41e8f64a9ae2b49b3c66e8f1b9b3940b390f2a934974073d6e4afee137"} err="failed to get container status \"e5acfe41e8f64a9ae2b49b3c66e8f1b9b3940b390f2a934974073d6e4afee137\": rpc error: code = NotFound desc = could not find container \"e5acfe41e8f64a9ae2b49b3c66e8f1b9b3940b390f2a934974073d6e4afee137\": container with ID starting with e5acfe41e8f64a9ae2b49b3c66e8f1b9b3940b390f2a934974073d6e4afee137 not found: ID does not exist" Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.134731 4828 scope.go:117] "RemoveContainer" containerID="f38682501ce035cb7c913628c69c4d62174d66e1903f367119c8168eac7d4095" Dec 10 19:00:57 crc kubenswrapper[4828]: E1210 19:00:57.135205 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f38682501ce035cb7c913628c69c4d62174d66e1903f367119c8168eac7d4095\": container with ID starting with f38682501ce035cb7c913628c69c4d62174d66e1903f367119c8168eac7d4095 not found: ID does not exist" containerID="f38682501ce035cb7c913628c69c4d62174d66e1903f367119c8168eac7d4095" Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.135234 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f38682501ce035cb7c913628c69c4d62174d66e1903f367119c8168eac7d4095"} err="failed to get container status \"f38682501ce035cb7c913628c69c4d62174d66e1903f367119c8168eac7d4095\": rpc error: code = NotFound desc = could not find container \"f38682501ce035cb7c913628c69c4d62174d66e1903f367119c8168eac7d4095\": container with ID starting with f38682501ce035cb7c913628c69c4d62174d66e1903f367119c8168eac7d4095 not found: ID does not exist" Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.135286 4828 scope.go:117] "RemoveContainer" containerID="32acb243cd59bea695a829412e491cd934342d0e3457c6e66993db1cc091c887" Dec 10 19:00:57 crc kubenswrapper[4828]: E1210 19:00:57.135657 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32acb243cd59bea695a829412e491cd934342d0e3457c6e66993db1cc091c887\": container with ID starting with 32acb243cd59bea695a829412e491cd934342d0e3457c6e66993db1cc091c887 not found: ID does not exist" containerID="32acb243cd59bea695a829412e491cd934342d0e3457c6e66993db1cc091c887" Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.135685 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32acb243cd59bea695a829412e491cd934342d0e3457c6e66993db1cc091c887"} err="failed to get container status \"32acb243cd59bea695a829412e491cd934342d0e3457c6e66993db1cc091c887\": rpc error: code = NotFound desc = could not find container \"32acb243cd59bea695a829412e491cd934342d0e3457c6e66993db1cc091c887\": container with ID starting with 32acb243cd59bea695a829412e491cd934342d0e3457c6e66993db1cc091c887 not found: ID does not exist" Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.803328 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" path="/var/lib/kubelet/pods/3ef84f18-e802-4038-ba08-2b4eb948d803/volumes" Dec 10 19:00:57 crc kubenswrapper[4828]: I1210 19:00:57.804701 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" path="/var/lib/kubelet/pods/e466f0bb-f597-46e1-b27a-a361322a6dcd/volumes" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.592525 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-5f59z"] Dec 10 19:01:02 crc kubenswrapper[4828]: E1210 19:01:02.593362 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" containerName="extract-utilities" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593379 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" containerName="extract-utilities" Dec 10 19:01:02 crc kubenswrapper[4828]: E1210 19:01:02.593392 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" containerName="extract-utilities" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593400 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" containerName="extract-utilities" Dec 10 19:01:02 crc kubenswrapper[4828]: E1210 19:01:02.593409 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" containerName="extract-utilities" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593418 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" containerName="extract-utilities" Dec 10 19:01:02 crc kubenswrapper[4828]: E1210 19:01:02.593430 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" containerName="registry-server" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593437 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" containerName="registry-server" Dec 10 19:01:02 crc kubenswrapper[4828]: E1210 19:01:02.593447 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" containerName="registry-server" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593454 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" containerName="registry-server" Dec 10 19:01:02 crc kubenswrapper[4828]: E1210 19:01:02.593467 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" containerName="registry-server" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593476 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" containerName="registry-server" Dec 10 19:01:02 crc kubenswrapper[4828]: E1210 19:01:02.593487 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22097d8d-0d50-4948-984b-4ea71356df75" containerName="collect-profiles" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593495 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="22097d8d-0d50-4948-984b-4ea71356df75" containerName="collect-profiles" Dec 10 19:01:02 crc kubenswrapper[4828]: E1210 19:01:02.593509 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" containerName="extract-utilities" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593516 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" containerName="extract-utilities" Dec 10 19:01:02 crc kubenswrapper[4828]: E1210 19:01:02.593525 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" containerName="extract-content" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593533 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" containerName="extract-content" Dec 10 19:01:02 crc kubenswrapper[4828]: E1210 19:01:02.593546 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" containerName="extract-content" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593552 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" containerName="extract-content" Dec 10 19:01:02 crc kubenswrapper[4828]: E1210 19:01:02.593563 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" containerName="extract-content" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593570 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" containerName="extract-content" Dec 10 19:01:02 crc kubenswrapper[4828]: E1210 19:01:02.593578 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" containerName="extract-content" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593586 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" containerName="extract-content" Dec 10 19:01:02 crc kubenswrapper[4828]: E1210 19:01:02.593595 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" containerName="registry-server" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593602 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" containerName="registry-server" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593701 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="e466f0bb-f597-46e1-b27a-a361322a6dcd" containerName="registry-server" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593715 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="44b88622-40eb-4bcf-81a5-dab1f8b4c63b" containerName="registry-server" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593728 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ef84f18-e802-4038-ba08-2b4eb948d803" containerName="registry-server" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593740 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="43cd3400-d5e9-4279-9a85-051625ca4607" containerName="registry-server" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.593751 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="22097d8d-0d50-4948-984b-4ea71356df75" containerName="collect-profiles" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.594273 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.604960 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-5f59z"] Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.655077 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.655154 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.655194 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.655228 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-registry-certificates\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.655255 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-registry-tls\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.655306 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrqf8\" (UniqueName: \"kubernetes.io/projected/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-kube-api-access-jrqf8\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.655332 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-trusted-ca\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.655365 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-bound-sa-token\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.683086 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.757435 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.757544 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-registry-certificates\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.757585 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-registry-tls\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.757624 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-trusted-ca\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.757657 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrqf8\" (UniqueName: \"kubernetes.io/projected/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-kube-api-access-jrqf8\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.757691 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-bound-sa-token\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.757784 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.758966 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-registry-certificates\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.759398 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-trusted-ca\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.759857 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.764123 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.765250 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-registry-tls\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.774886 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-bound-sa-token\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.775272 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrqf8\" (UniqueName: \"kubernetes.io/projected/eb9b8754-fabb-4bef-87c4-5ef6314b7b1e-kube-api-access-jrqf8\") pod \"image-registry-66df7c8f76-5f59z\" (UID: \"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e\") " pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:02 crc kubenswrapper[4828]: I1210 19:01:02.910502 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:03 crc kubenswrapper[4828]: I1210 19:01:03.109612 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-5f59z"] Dec 10 19:01:04 crc kubenswrapper[4828]: I1210 19:01:04.056679 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" event={"ID":"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e","Type":"ContainerStarted","Data":"75faee1e3ac8ecf5fff08a8e16ac34b8e87d39edf58e9f38774693baf356a3e2"} Dec 10 19:01:04 crc kubenswrapper[4828]: I1210 19:01:04.057046 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" event={"ID":"eb9b8754-fabb-4bef-87c4-5ef6314b7b1e","Type":"ContainerStarted","Data":"192cdaa7d82cb37ff3712a32139bddf69b631845622fc71c8ad7e34529aea028"} Dec 10 19:01:04 crc kubenswrapper[4828]: I1210 19:01:04.058136 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:04 crc kubenswrapper[4828]: I1210 19:01:04.085455 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" podStartSLOduration=2.085432331 podStartE2EDuration="2.085432331s" podCreationTimestamp="2025-12-10 19:01:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:01:04.082670638 +0000 UTC m=+344.593281683" watchObservedRunningTime="2025-12-10 19:01:04.085432331 +0000 UTC m=+344.596043346" Dec 10 19:01:21 crc kubenswrapper[4828]: I1210 19:01:21.231158 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:01:21 crc kubenswrapper[4828]: I1210 19:01:21.232001 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:01:22 crc kubenswrapper[4828]: I1210 19:01:22.919285 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-5f59z" Dec 10 19:01:22 crc kubenswrapper[4828]: I1210 19:01:22.973957 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c4wpt"] Dec 10 19:01:31 crc kubenswrapper[4828]: I1210 19:01:31.963265 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gs2g5"] Dec 10 19:01:31 crc kubenswrapper[4828]: I1210 19:01:31.964192 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gs2g5" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" containerName="registry-server" containerID="cri-o://42e14c173fa2164db28101ad65272161516c6ce98e5793101e69699ba00e4747" gracePeriod=30 Dec 10 19:01:31 crc kubenswrapper[4828]: I1210 19:01:31.969598 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7xlcc"] Dec 10 19:01:31 crc kubenswrapper[4828]: I1210 19:01:31.969927 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7xlcc" podUID="1f95c086-c5d6-4ad2-97e6-79e652fa0040" containerName="registry-server" containerID="cri-o://43ea3e48374dc6c2aff5c489ad72c1d2572057d86cbf9d439754899b12f03383" gracePeriod=30 Dec 10 19:01:31 crc kubenswrapper[4828]: I1210 19:01:31.981326 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w7n9w"] Dec 10 19:01:31 crc kubenswrapper[4828]: I1210 19:01:31.981998 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" podUID="c5bd6649-abb0-4898-a73f-efceb62d2cb9" containerName="marketplace-operator" containerID="cri-o://268da9959d149cddc863357066862e37a4f201fdd5fda2fc8175ac3266496ac9" gracePeriod=30 Dec 10 19:01:31 crc kubenswrapper[4828]: I1210 19:01:31.987721 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qs5qt"] Dec 10 19:01:31 crc kubenswrapper[4828]: I1210 19:01:31.988019 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qs5qt" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" containerName="registry-server" containerID="cri-o://b32875c68b7888406fadd752847829bdc52c11da3f660b23e7a1d0d41c6f2c3b" gracePeriod=30 Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.004052 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-f74sl"] Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.004834 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-f74sl" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.007633 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5d8xc"] Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.013657 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-f74sl"] Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.019003 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5d8xc" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" containerName="registry-server" containerID="cri-o://db2abf9b0f92038dc8c11083bf5dabea39754b8c6e01f571bf1429744d72c324" gracePeriod=30 Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.161725 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q58n2\" (UniqueName: \"kubernetes.io/projected/8df288f1-1cd4-4381-b519-5a353cd63c59-kube-api-access-q58n2\") pod \"marketplace-operator-79b997595-f74sl\" (UID: \"8df288f1-1cd4-4381-b519-5a353cd63c59\") " pod="openshift-marketplace/marketplace-operator-79b997595-f74sl" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.161848 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8df288f1-1cd4-4381-b519-5a353cd63c59-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-f74sl\" (UID: \"8df288f1-1cd4-4381-b519-5a353cd63c59\") " pod="openshift-marketplace/marketplace-operator-79b997595-f74sl" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.161879 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8df288f1-1cd4-4381-b519-5a353cd63c59-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-f74sl\" (UID: \"8df288f1-1cd4-4381-b519-5a353cd63c59\") " pod="openshift-marketplace/marketplace-operator-79b997595-f74sl" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.233222 4828 generic.go:334] "Generic (PLEG): container finished" podID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" containerID="b32875c68b7888406fadd752847829bdc52c11da3f660b23e7a1d0d41c6f2c3b" exitCode=0 Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.233301 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs5qt" event={"ID":"b6084fc6-1311-4715-9d6d-01b42fec5f4b","Type":"ContainerDied","Data":"b32875c68b7888406fadd752847829bdc52c11da3f660b23e7a1d0d41c6f2c3b"} Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.235690 4828 generic.go:334] "Generic (PLEG): container finished" podID="c5bd6649-abb0-4898-a73f-efceb62d2cb9" containerID="268da9959d149cddc863357066862e37a4f201fdd5fda2fc8175ac3266496ac9" exitCode=0 Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.235741 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" event={"ID":"c5bd6649-abb0-4898-a73f-efceb62d2cb9","Type":"ContainerDied","Data":"268da9959d149cddc863357066862e37a4f201fdd5fda2fc8175ac3266496ac9"} Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.235779 4828 scope.go:117] "RemoveContainer" containerID="2d897701418eef8ce32adfb032b1a0b184b7fe03ec55abcb747335b88b3ebb7d" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.240038 4828 generic.go:334] "Generic (PLEG): container finished" podID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" containerID="42e14c173fa2164db28101ad65272161516c6ce98e5793101e69699ba00e4747" exitCode=0 Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.240083 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gs2g5" event={"ID":"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc","Type":"ContainerDied","Data":"42e14c173fa2164db28101ad65272161516c6ce98e5793101e69699ba00e4747"} Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.243023 4828 generic.go:334] "Generic (PLEG): container finished" podID="2e51cf27-f229-4c52-b839-c5f957e0fa99" containerID="db2abf9b0f92038dc8c11083bf5dabea39754b8c6e01f571bf1429744d72c324" exitCode=0 Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.243074 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5d8xc" event={"ID":"2e51cf27-f229-4c52-b839-c5f957e0fa99","Type":"ContainerDied","Data":"db2abf9b0f92038dc8c11083bf5dabea39754b8c6e01f571bf1429744d72c324"} Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.245357 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f95c086-c5d6-4ad2-97e6-79e652fa0040" containerID="43ea3e48374dc6c2aff5c489ad72c1d2572057d86cbf9d439754899b12f03383" exitCode=0 Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.245380 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xlcc" event={"ID":"1f95c086-c5d6-4ad2-97e6-79e652fa0040","Type":"ContainerDied","Data":"43ea3e48374dc6c2aff5c489ad72c1d2572057d86cbf9d439754899b12f03383"} Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.262566 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8df288f1-1cd4-4381-b519-5a353cd63c59-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-f74sl\" (UID: \"8df288f1-1cd4-4381-b519-5a353cd63c59\") " pod="openshift-marketplace/marketplace-operator-79b997595-f74sl" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.262611 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8df288f1-1cd4-4381-b519-5a353cd63c59-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-f74sl\" (UID: \"8df288f1-1cd4-4381-b519-5a353cd63c59\") " pod="openshift-marketplace/marketplace-operator-79b997595-f74sl" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.262631 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q58n2\" (UniqueName: \"kubernetes.io/projected/8df288f1-1cd4-4381-b519-5a353cd63c59-kube-api-access-q58n2\") pod \"marketplace-operator-79b997595-f74sl\" (UID: \"8df288f1-1cd4-4381-b519-5a353cd63c59\") " pod="openshift-marketplace/marketplace-operator-79b997595-f74sl" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.264683 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8df288f1-1cd4-4381-b519-5a353cd63c59-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-f74sl\" (UID: \"8df288f1-1cd4-4381-b519-5a353cd63c59\") " pod="openshift-marketplace/marketplace-operator-79b997595-f74sl" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.269651 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8df288f1-1cd4-4381-b519-5a353cd63c59-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-f74sl\" (UID: \"8df288f1-1cd4-4381-b519-5a353cd63c59\") " pod="openshift-marketplace/marketplace-operator-79b997595-f74sl" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.278409 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q58n2\" (UniqueName: \"kubernetes.io/projected/8df288f1-1cd4-4381-b519-5a353cd63c59-kube-api-access-q58n2\") pod \"marketplace-operator-79b997595-f74sl\" (UID: \"8df288f1-1cd4-4381-b519-5a353cd63c59\") " pod="openshift-marketplace/marketplace-operator-79b997595-f74sl" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.321327 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-f74sl" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.480390 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gs2g5" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.483263 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5d8xc" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.491665 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7xlcc" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.507018 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.508766 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qs5qt" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.566042 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f95c086-c5d6-4ad2-97e6-79e652fa0040-utilities\") pod \"1f95c086-c5d6-4ad2-97e6-79e652fa0040\" (UID: \"1f95c086-c5d6-4ad2-97e6-79e652fa0040\") " Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.566116 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-utilities\") pod \"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc\" (UID: \"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc\") " Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.566144 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e51cf27-f229-4c52-b839-c5f957e0fa99-utilities\") pod \"2e51cf27-f229-4c52-b839-c5f957e0fa99\" (UID: \"2e51cf27-f229-4c52-b839-c5f957e0fa99\") " Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.566167 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmntb\" (UniqueName: \"kubernetes.io/projected/1f95c086-c5d6-4ad2-97e6-79e652fa0040-kube-api-access-vmntb\") pod \"1f95c086-c5d6-4ad2-97e6-79e652fa0040\" (UID: \"1f95c086-c5d6-4ad2-97e6-79e652fa0040\") " Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.566192 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxbj6\" (UniqueName: \"kubernetes.io/projected/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-kube-api-access-rxbj6\") pod \"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc\" (UID: \"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc\") " Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.566214 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f95c086-c5d6-4ad2-97e6-79e652fa0040-catalog-content\") pod \"1f95c086-c5d6-4ad2-97e6-79e652fa0040\" (UID: \"1f95c086-c5d6-4ad2-97e6-79e652fa0040\") " Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.566233 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-catalog-content\") pod \"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc\" (UID: \"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc\") " Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.566251 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e51cf27-f229-4c52-b839-c5f957e0fa99-catalog-content\") pod \"2e51cf27-f229-4c52-b839-c5f957e0fa99\" (UID: \"2e51cf27-f229-4c52-b839-c5f957e0fa99\") " Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.566313 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpl52\" (UniqueName: \"kubernetes.io/projected/2e51cf27-f229-4c52-b839-c5f957e0fa99-kube-api-access-qpl52\") pod \"2e51cf27-f229-4c52-b839-c5f957e0fa99\" (UID: \"2e51cf27-f229-4c52-b839-c5f957e0fa99\") " Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.568320 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-utilities" (OuterVolumeSpecName: "utilities") pod "0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" (UID: "0dbff95a-c195-4a72-bbfe-0a31a3feb8cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.569215 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f95c086-c5d6-4ad2-97e6-79e652fa0040-utilities" (OuterVolumeSpecName: "utilities") pod "1f95c086-c5d6-4ad2-97e6-79e652fa0040" (UID: "1f95c086-c5d6-4ad2-97e6-79e652fa0040"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.569993 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e51cf27-f229-4c52-b839-c5f957e0fa99-utilities" (OuterVolumeSpecName: "utilities") pod "2e51cf27-f229-4c52-b839-c5f957e0fa99" (UID: "2e51cf27-f229-4c52-b839-c5f957e0fa99"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.574245 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f95c086-c5d6-4ad2-97e6-79e652fa0040-kube-api-access-vmntb" (OuterVolumeSpecName: "kube-api-access-vmntb") pod "1f95c086-c5d6-4ad2-97e6-79e652fa0040" (UID: "1f95c086-c5d6-4ad2-97e6-79e652fa0040"). InnerVolumeSpecName "kube-api-access-vmntb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.576496 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e51cf27-f229-4c52-b839-c5f957e0fa99-kube-api-access-qpl52" (OuterVolumeSpecName: "kube-api-access-qpl52") pod "2e51cf27-f229-4c52-b839-c5f957e0fa99" (UID: "2e51cf27-f229-4c52-b839-c5f957e0fa99"). InnerVolumeSpecName "kube-api-access-qpl52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.578529 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-kube-api-access-rxbj6" (OuterVolumeSpecName: "kube-api-access-rxbj6") pod "0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" (UID: "0dbff95a-c195-4a72-bbfe-0a31a3feb8cc"). InnerVolumeSpecName "kube-api-access-rxbj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.628995 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" (UID: "0dbff95a-c195-4a72-bbfe-0a31a3feb8cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.629576 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f95c086-c5d6-4ad2-97e6-79e652fa0040-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1f95c086-c5d6-4ad2-97e6-79e652fa0040" (UID: "1f95c086-c5d6-4ad2-97e6-79e652fa0040"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.667679 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6084fc6-1311-4715-9d6d-01b42fec5f4b-utilities\") pod \"b6084fc6-1311-4715-9d6d-01b42fec5f4b\" (UID: \"b6084fc6-1311-4715-9d6d-01b42fec5f4b\") " Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.667760 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9t2q\" (UniqueName: \"kubernetes.io/projected/b6084fc6-1311-4715-9d6d-01b42fec5f4b-kube-api-access-n9t2q\") pod \"b6084fc6-1311-4715-9d6d-01b42fec5f4b\" (UID: \"b6084fc6-1311-4715-9d6d-01b42fec5f4b\") " Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.667785 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c5bd6649-abb0-4898-a73f-efceb62d2cb9-marketplace-operator-metrics\") pod \"c5bd6649-abb0-4898-a73f-efceb62d2cb9\" (UID: \"c5bd6649-abb0-4898-a73f-efceb62d2cb9\") " Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.667885 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6084fc6-1311-4715-9d6d-01b42fec5f4b-catalog-content\") pod \"b6084fc6-1311-4715-9d6d-01b42fec5f4b\" (UID: \"b6084fc6-1311-4715-9d6d-01b42fec5f4b\") " Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.667936 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5bd6649-abb0-4898-a73f-efceb62d2cb9-marketplace-trusted-ca\") pod \"c5bd6649-abb0-4898-a73f-efceb62d2cb9\" (UID: \"c5bd6649-abb0-4898-a73f-efceb62d2cb9\") " Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.668000 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qvwq\" (UniqueName: \"kubernetes.io/projected/c5bd6649-abb0-4898-a73f-efceb62d2cb9-kube-api-access-4qvwq\") pod \"c5bd6649-abb0-4898-a73f-efceb62d2cb9\" (UID: \"c5bd6649-abb0-4898-a73f-efceb62d2cb9\") " Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.668278 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxbj6\" (UniqueName: \"kubernetes.io/projected/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-kube-api-access-rxbj6\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.668314 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.668323 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1f95c086-c5d6-4ad2-97e6-79e652fa0040-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.668333 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpl52\" (UniqueName: \"kubernetes.io/projected/2e51cf27-f229-4c52-b839-c5f957e0fa99-kube-api-access-qpl52\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.668342 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1f95c086-c5d6-4ad2-97e6-79e652fa0040-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.668351 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.668377 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e51cf27-f229-4c52-b839-c5f957e0fa99-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.668387 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmntb\" (UniqueName: \"kubernetes.io/projected/1f95c086-c5d6-4ad2-97e6-79e652fa0040-kube-api-access-vmntb\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.669686 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6084fc6-1311-4715-9d6d-01b42fec5f4b-utilities" (OuterVolumeSpecName: "utilities") pod "b6084fc6-1311-4715-9d6d-01b42fec5f4b" (UID: "b6084fc6-1311-4715-9d6d-01b42fec5f4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.669934 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5bd6649-abb0-4898-a73f-efceb62d2cb9-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "c5bd6649-abb0-4898-a73f-efceb62d2cb9" (UID: "c5bd6649-abb0-4898-a73f-efceb62d2cb9"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.671074 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5bd6649-abb0-4898-a73f-efceb62d2cb9-kube-api-access-4qvwq" (OuterVolumeSpecName: "kube-api-access-4qvwq") pod "c5bd6649-abb0-4898-a73f-efceb62d2cb9" (UID: "c5bd6649-abb0-4898-a73f-efceb62d2cb9"). InnerVolumeSpecName "kube-api-access-4qvwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.671270 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5bd6649-abb0-4898-a73f-efceb62d2cb9-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "c5bd6649-abb0-4898-a73f-efceb62d2cb9" (UID: "c5bd6649-abb0-4898-a73f-efceb62d2cb9"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.672005 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6084fc6-1311-4715-9d6d-01b42fec5f4b-kube-api-access-n9t2q" (OuterVolumeSpecName: "kube-api-access-n9t2q") pod "b6084fc6-1311-4715-9d6d-01b42fec5f4b" (UID: "b6084fc6-1311-4715-9d6d-01b42fec5f4b"). InnerVolumeSpecName "kube-api-access-n9t2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.687978 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e51cf27-f229-4c52-b839-c5f957e0fa99-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e51cf27-f229-4c52-b839-c5f957e0fa99" (UID: "2e51cf27-f229-4c52-b839-c5f957e0fa99"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.690288 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6084fc6-1311-4715-9d6d-01b42fec5f4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b6084fc6-1311-4715-9d6d-01b42fec5f4b" (UID: "b6084fc6-1311-4715-9d6d-01b42fec5f4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.758659 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-f74sl"] Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.769050 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qvwq\" (UniqueName: \"kubernetes.io/projected/c5bd6649-abb0-4898-a73f-efceb62d2cb9-kube-api-access-4qvwq\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.769080 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6084fc6-1311-4715-9d6d-01b42fec5f4b-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.769093 4828 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c5bd6649-abb0-4898-a73f-efceb62d2cb9-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.769106 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9t2q\" (UniqueName: \"kubernetes.io/projected/b6084fc6-1311-4715-9d6d-01b42fec5f4b-kube-api-access-n9t2q\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.769116 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e51cf27-f229-4c52-b839-c5f957e0fa99-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.769127 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6084fc6-1311-4715-9d6d-01b42fec5f4b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:32 crc kubenswrapper[4828]: I1210 19:01:32.769139 4828 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c5bd6649-abb0-4898-a73f-efceb62d2cb9-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.253359 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gs2g5" event={"ID":"0dbff95a-c195-4a72-bbfe-0a31a3feb8cc","Type":"ContainerDied","Data":"07cf7a4a2d61c5484c036d16fb34f3779c57e1774cec637701262cdd09edaf90"} Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.253887 4828 scope.go:117] "RemoveContainer" containerID="42e14c173fa2164db28101ad65272161516c6ce98e5793101e69699ba00e4747" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.253436 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gs2g5" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.260058 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7xlcc" event={"ID":"1f95c086-c5d6-4ad2-97e6-79e652fa0040","Type":"ContainerDied","Data":"3e9a991f46cc58755029cd47d9c6053f6d65187a531e9540fef390e2a5013f6b"} Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.260194 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7xlcc" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.265255 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5d8xc" event={"ID":"2e51cf27-f229-4c52-b839-c5f957e0fa99","Type":"ContainerDied","Data":"e3d1ba51ed316152866acef073e5d0d5ee7fbd13299905e3805dc065ffc2c0e6"} Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.265308 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5d8xc" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.267458 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-f74sl" event={"ID":"8df288f1-1cd4-4381-b519-5a353cd63c59","Type":"ContainerStarted","Data":"f8a3430e5f1c3a4f89af630a6839ecfe8a494e915675ea4d765a523855e2cc05"} Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.267571 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-f74sl" event={"ID":"8df288f1-1cd4-4381-b519-5a353cd63c59","Type":"ContainerStarted","Data":"d933f703455569432a15e8bb8e484266df7dda28c97133ac7e8887bbcc597438"} Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.268572 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-f74sl" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.271071 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qs5qt" event={"ID":"b6084fc6-1311-4715-9d6d-01b42fec5f4b","Type":"ContainerDied","Data":"6f482adc183180d7a611457a1dca9793deb7270bb71879b9563741f65c1c9c39"} Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.271299 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qs5qt" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.274538 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-f74sl" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.276644 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" event={"ID":"c5bd6649-abb0-4898-a73f-efceb62d2cb9","Type":"ContainerDied","Data":"99f36d3821e1a5847f353ca35068711becb42bfec67bef0fcffed7947792b39e"} Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.276829 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-w7n9w" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.277976 4828 scope.go:117] "RemoveContainer" containerID="ae1236337f8725a6a27cfcff66f4460525b8ab1833f64f3e8880efd3d63a8aef" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.285394 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-f74sl" podStartSLOduration=2.285378184 podStartE2EDuration="2.285378184s" podCreationTimestamp="2025-12-10 19:01:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:01:33.284452366 +0000 UTC m=+373.795063381" watchObservedRunningTime="2025-12-10 19:01:33.285378184 +0000 UTC m=+373.795989199" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.305633 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gs2g5"] Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.309550 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gs2g5"] Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.329889 4828 scope.go:117] "RemoveContainer" containerID="bbbaacb63de04ed7a04804f53dcde560ec5ac83aa5ab751bf45b94e6876c57cb" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.400127 4828 scope.go:117] "RemoveContainer" containerID="43ea3e48374dc6c2aff5c489ad72c1d2572057d86cbf9d439754899b12f03383" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.409246 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5d8xc"] Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.416568 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5d8xc"] Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.419575 4828 scope.go:117] "RemoveContainer" containerID="f62c9e024dd01b2a6750729b340a9a88f3da3ca61af4f0eb2db395d87ebe232c" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.434092 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7xlcc"] Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.436080 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7xlcc"] Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.437977 4828 scope.go:117] "RemoveContainer" containerID="b036d88617a44b0d9fdb9d1570cb2a10a128f97f9fb9e308da23fb200a861ad7" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.445009 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w7n9w"] Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.451256 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-w7n9w"] Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.456481 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qs5qt"] Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.457311 4828 scope.go:117] "RemoveContainer" containerID="db2abf9b0f92038dc8c11083bf5dabea39754b8c6e01f571bf1429744d72c324" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.459071 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qs5qt"] Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.470754 4828 scope.go:117] "RemoveContainer" containerID="d92f5cf24f955edfd9ed3b157288034cdce48f76a3a5d17afb4c89f60f80d14d" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.482350 4828 scope.go:117] "RemoveContainer" containerID="b9a195a31ee8a73c0f3b8191ce4e21ba46ebaff3eb3780edaa59f7dd4eb56473" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.497047 4828 scope.go:117] "RemoveContainer" containerID="b32875c68b7888406fadd752847829bdc52c11da3f660b23e7a1d0d41c6f2c3b" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.510339 4828 scope.go:117] "RemoveContainer" containerID="5779b1692ba90a6ea89eeced9e7d6749c00e508dc24744471fe2efd9ed8fcaf0" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.522626 4828 scope.go:117] "RemoveContainer" containerID="92c81df57e70189d396ba6ff6646f0c857cb35e8e37710822d1eafb5702214a8" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.533185 4828 scope.go:117] "RemoveContainer" containerID="268da9959d149cddc863357066862e37a4f201fdd5fda2fc8175ac3266496ac9" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.794962 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" path="/var/lib/kubelet/pods/0dbff95a-c195-4a72-bbfe-0a31a3feb8cc/volumes" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.795860 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f95c086-c5d6-4ad2-97e6-79e652fa0040" path="/var/lib/kubelet/pods/1f95c086-c5d6-4ad2-97e6-79e652fa0040/volumes" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.796722 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" path="/var/lib/kubelet/pods/2e51cf27-f229-4c52-b839-c5f957e0fa99/volumes" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.798102 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" path="/var/lib/kubelet/pods/b6084fc6-1311-4715-9d6d-01b42fec5f4b/volumes" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.798957 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5bd6649-abb0-4898-a73f-efceb62d2cb9" path="/var/lib/kubelet/pods/c5bd6649-abb0-4898-a73f-efceb62d2cb9/volumes" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.972308 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-df2d5"] Dec 10 19:01:33 crc kubenswrapper[4828]: E1210 19:01:33.972837 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" containerName="extract-utilities" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.972848 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" containerName="extract-utilities" Dec 10 19:01:33 crc kubenswrapper[4828]: E1210 19:01:33.972858 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" containerName="extract-utilities" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.972864 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" containerName="extract-utilities" Dec 10 19:01:33 crc kubenswrapper[4828]: E1210 19:01:33.972873 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f95c086-c5d6-4ad2-97e6-79e652fa0040" containerName="extract-content" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.972879 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f95c086-c5d6-4ad2-97e6-79e652fa0040" containerName="extract-content" Dec 10 19:01:33 crc kubenswrapper[4828]: E1210 19:01:33.972887 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" containerName="registry-server" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.972892 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" containerName="registry-server" Dec 10 19:01:33 crc kubenswrapper[4828]: E1210 19:01:33.972898 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" containerName="registry-server" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.972903 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" containerName="registry-server" Dec 10 19:01:33 crc kubenswrapper[4828]: E1210 19:01:33.972911 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f95c086-c5d6-4ad2-97e6-79e652fa0040" containerName="extract-utilities" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.972916 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f95c086-c5d6-4ad2-97e6-79e652fa0040" containerName="extract-utilities" Dec 10 19:01:33 crc kubenswrapper[4828]: E1210 19:01:33.972924 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5bd6649-abb0-4898-a73f-efceb62d2cb9" containerName="marketplace-operator" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.972929 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5bd6649-abb0-4898-a73f-efceb62d2cb9" containerName="marketplace-operator" Dec 10 19:01:33 crc kubenswrapper[4828]: E1210 19:01:33.972941 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" containerName="registry-server" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.972946 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" containerName="registry-server" Dec 10 19:01:33 crc kubenswrapper[4828]: E1210 19:01:33.972957 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" containerName="extract-content" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.972963 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" containerName="extract-content" Dec 10 19:01:33 crc kubenswrapper[4828]: E1210 19:01:33.972970 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" containerName="extract-content" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.972975 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" containerName="extract-content" Dec 10 19:01:33 crc kubenswrapper[4828]: E1210 19:01:33.972981 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" containerName="extract-utilities" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.972986 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" containerName="extract-utilities" Dec 10 19:01:33 crc kubenswrapper[4828]: E1210 19:01:33.972994 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" containerName="extract-content" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.972999 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" containerName="extract-content" Dec 10 19:01:33 crc kubenswrapper[4828]: E1210 19:01:33.973007 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f95c086-c5d6-4ad2-97e6-79e652fa0040" containerName="registry-server" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.973013 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f95c086-c5d6-4ad2-97e6-79e652fa0040" containerName="registry-server" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.973092 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e51cf27-f229-4c52-b839-c5f957e0fa99" containerName="registry-server" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.973101 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5bd6649-abb0-4898-a73f-efceb62d2cb9" containerName="marketplace-operator" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.973110 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6084fc6-1311-4715-9d6d-01b42fec5f4b" containerName="registry-server" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.973119 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5bd6649-abb0-4898-a73f-efceb62d2cb9" containerName="marketplace-operator" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.973129 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f95c086-c5d6-4ad2-97e6-79e652fa0040" containerName="registry-server" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.973136 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dbff95a-c195-4a72-bbfe-0a31a3feb8cc" containerName="registry-server" Dec 10 19:01:33 crc kubenswrapper[4828]: E1210 19:01:33.973213 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5bd6649-abb0-4898-a73f-efceb62d2cb9" containerName="marketplace-operator" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.973220 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5bd6649-abb0-4898-a73f-efceb62d2cb9" containerName="marketplace-operator" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.973765 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-df2d5" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.976418 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 10 19:01:33 crc kubenswrapper[4828]: I1210 19:01:33.985724 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-df2d5"] Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.087145 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjv67\" (UniqueName: \"kubernetes.io/projected/fb55e51a-e3e7-4bea-8791-aec28f2b8873-kube-api-access-jjv67\") pod \"certified-operators-df2d5\" (UID: \"fb55e51a-e3e7-4bea-8791-aec28f2b8873\") " pod="openshift-marketplace/certified-operators-df2d5" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.087379 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb55e51a-e3e7-4bea-8791-aec28f2b8873-utilities\") pod \"certified-operators-df2d5\" (UID: \"fb55e51a-e3e7-4bea-8791-aec28f2b8873\") " pod="openshift-marketplace/certified-operators-df2d5" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.087477 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb55e51a-e3e7-4bea-8791-aec28f2b8873-catalog-content\") pod \"certified-operators-df2d5\" (UID: \"fb55e51a-e3e7-4bea-8791-aec28f2b8873\") " pod="openshift-marketplace/certified-operators-df2d5" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.188212 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb55e51a-e3e7-4bea-8791-aec28f2b8873-catalog-content\") pod \"certified-operators-df2d5\" (UID: \"fb55e51a-e3e7-4bea-8791-aec28f2b8873\") " pod="openshift-marketplace/certified-operators-df2d5" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.188521 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjv67\" (UniqueName: \"kubernetes.io/projected/fb55e51a-e3e7-4bea-8791-aec28f2b8873-kube-api-access-jjv67\") pod \"certified-operators-df2d5\" (UID: \"fb55e51a-e3e7-4bea-8791-aec28f2b8873\") " pod="openshift-marketplace/certified-operators-df2d5" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.188680 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb55e51a-e3e7-4bea-8791-aec28f2b8873-utilities\") pod \"certified-operators-df2d5\" (UID: \"fb55e51a-e3e7-4bea-8791-aec28f2b8873\") " pod="openshift-marketplace/certified-operators-df2d5" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.188706 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb55e51a-e3e7-4bea-8791-aec28f2b8873-catalog-content\") pod \"certified-operators-df2d5\" (UID: \"fb55e51a-e3e7-4bea-8791-aec28f2b8873\") " pod="openshift-marketplace/certified-operators-df2d5" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.188928 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb55e51a-e3e7-4bea-8791-aec28f2b8873-utilities\") pod \"certified-operators-df2d5\" (UID: \"fb55e51a-e3e7-4bea-8791-aec28f2b8873\") " pod="openshift-marketplace/certified-operators-df2d5" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.206334 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjv67\" (UniqueName: \"kubernetes.io/projected/fb55e51a-e3e7-4bea-8791-aec28f2b8873-kube-api-access-jjv67\") pod \"certified-operators-df2d5\" (UID: \"fb55e51a-e3e7-4bea-8791-aec28f2b8873\") " pod="openshift-marketplace/certified-operators-df2d5" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.291306 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-df2d5" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.492399 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-df2d5"] Dec 10 19:01:34 crc kubenswrapper[4828]: W1210 19:01:34.496622 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb55e51a_e3e7_4bea_8791_aec28f2b8873.slice/crio-f70d4dc20b82d206c633243e4b1c73012bb8e906b5b227311682ff1bd9e10be2 WatchSource:0}: Error finding container f70d4dc20b82d206c633243e4b1c73012bb8e906b5b227311682ff1bd9e10be2: Status 404 returned error can't find the container with id f70d4dc20b82d206c633243e4b1c73012bb8e906b5b227311682ff1bd9e10be2 Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.576910 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hm6hf"] Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.577840 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hm6hf" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.580204 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.587494 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hm6hf"] Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.696966 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b49ed428-86d9-4b1d-a2c6-f3f79fadb13d-utilities\") pod \"redhat-marketplace-hm6hf\" (UID: \"b49ed428-86d9-4b1d-a2c6-f3f79fadb13d\") " pod="openshift-marketplace/redhat-marketplace-hm6hf" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.697024 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b49ed428-86d9-4b1d-a2c6-f3f79fadb13d-catalog-content\") pod \"redhat-marketplace-hm6hf\" (UID: \"b49ed428-86d9-4b1d-a2c6-f3f79fadb13d\") " pod="openshift-marketplace/redhat-marketplace-hm6hf" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.697054 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg7pw\" (UniqueName: \"kubernetes.io/projected/b49ed428-86d9-4b1d-a2c6-f3f79fadb13d-kube-api-access-jg7pw\") pod \"redhat-marketplace-hm6hf\" (UID: \"b49ed428-86d9-4b1d-a2c6-f3f79fadb13d\") " pod="openshift-marketplace/redhat-marketplace-hm6hf" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.797930 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b49ed428-86d9-4b1d-a2c6-f3f79fadb13d-utilities\") pod \"redhat-marketplace-hm6hf\" (UID: \"b49ed428-86d9-4b1d-a2c6-f3f79fadb13d\") " pod="openshift-marketplace/redhat-marketplace-hm6hf" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.797987 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b49ed428-86d9-4b1d-a2c6-f3f79fadb13d-catalog-content\") pod \"redhat-marketplace-hm6hf\" (UID: \"b49ed428-86d9-4b1d-a2c6-f3f79fadb13d\") " pod="openshift-marketplace/redhat-marketplace-hm6hf" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.798018 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg7pw\" (UniqueName: \"kubernetes.io/projected/b49ed428-86d9-4b1d-a2c6-f3f79fadb13d-kube-api-access-jg7pw\") pod \"redhat-marketplace-hm6hf\" (UID: \"b49ed428-86d9-4b1d-a2c6-f3f79fadb13d\") " pod="openshift-marketplace/redhat-marketplace-hm6hf" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.798361 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b49ed428-86d9-4b1d-a2c6-f3f79fadb13d-utilities\") pod \"redhat-marketplace-hm6hf\" (UID: \"b49ed428-86d9-4b1d-a2c6-f3f79fadb13d\") " pod="openshift-marketplace/redhat-marketplace-hm6hf" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.798551 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b49ed428-86d9-4b1d-a2c6-f3f79fadb13d-catalog-content\") pod \"redhat-marketplace-hm6hf\" (UID: \"b49ed428-86d9-4b1d-a2c6-f3f79fadb13d\") " pod="openshift-marketplace/redhat-marketplace-hm6hf" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.826890 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg7pw\" (UniqueName: \"kubernetes.io/projected/b49ed428-86d9-4b1d-a2c6-f3f79fadb13d-kube-api-access-jg7pw\") pod \"redhat-marketplace-hm6hf\" (UID: \"b49ed428-86d9-4b1d-a2c6-f3f79fadb13d\") " pod="openshift-marketplace/redhat-marketplace-hm6hf" Dec 10 19:01:34 crc kubenswrapper[4828]: I1210 19:01:34.889735 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hm6hf" Dec 10 19:01:35 crc kubenswrapper[4828]: I1210 19:01:35.286925 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hm6hf"] Dec 10 19:01:35 crc kubenswrapper[4828]: W1210 19:01:35.298062 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb49ed428_86d9_4b1d_a2c6_f3f79fadb13d.slice/crio-1b818c8346c2ccbf354c60975f2caa48be446699435ad08fd2e363a31963606e WatchSource:0}: Error finding container 1b818c8346c2ccbf354c60975f2caa48be446699435ad08fd2e363a31963606e: Status 404 returned error can't find the container with id 1b818c8346c2ccbf354c60975f2caa48be446699435ad08fd2e363a31963606e Dec 10 19:01:35 crc kubenswrapper[4828]: I1210 19:01:35.317700 4828 generic.go:334] "Generic (PLEG): container finished" podID="fb55e51a-e3e7-4bea-8791-aec28f2b8873" containerID="480b31904be93ec6a5ebc33b518266280dbca275e4083c23f18ec3ddb6f94522" exitCode=0 Dec 10 19:01:35 crc kubenswrapper[4828]: I1210 19:01:35.317756 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-df2d5" event={"ID":"fb55e51a-e3e7-4bea-8791-aec28f2b8873","Type":"ContainerDied","Data":"480b31904be93ec6a5ebc33b518266280dbca275e4083c23f18ec3ddb6f94522"} Dec 10 19:01:35 crc kubenswrapper[4828]: I1210 19:01:35.317828 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-df2d5" event={"ID":"fb55e51a-e3e7-4bea-8791-aec28f2b8873","Type":"ContainerStarted","Data":"f70d4dc20b82d206c633243e4b1c73012bb8e906b5b227311682ff1bd9e10be2"} Dec 10 19:01:35 crc kubenswrapper[4828]: I1210 19:01:35.319831 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hm6hf" event={"ID":"b49ed428-86d9-4b1d-a2c6-f3f79fadb13d","Type":"ContainerStarted","Data":"1b818c8346c2ccbf354c60975f2caa48be446699435ad08fd2e363a31963606e"} Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.326443 4828 generic.go:334] "Generic (PLEG): container finished" podID="b49ed428-86d9-4b1d-a2c6-f3f79fadb13d" containerID="5aa3f43f769412944a71dbaf8af4dfaab6e3d285dcc70e01f3cecd11928a628a" exitCode=0 Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.326532 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hm6hf" event={"ID":"b49ed428-86d9-4b1d-a2c6-f3f79fadb13d","Type":"ContainerDied","Data":"5aa3f43f769412944a71dbaf8af4dfaab6e3d285dcc70e01f3cecd11928a628a"} Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.386199 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g7mhn"] Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.387345 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g7mhn" Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.390018 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.390077 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g7mhn"] Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.519145 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4vqt\" (UniqueName: \"kubernetes.io/projected/3a8395da-9d4b-4d62-be88-dd60993239ef-kube-api-access-l4vqt\") pod \"redhat-operators-g7mhn\" (UID: \"3a8395da-9d4b-4d62-be88-dd60993239ef\") " pod="openshift-marketplace/redhat-operators-g7mhn" Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.519194 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a8395da-9d4b-4d62-be88-dd60993239ef-catalog-content\") pod \"redhat-operators-g7mhn\" (UID: \"3a8395da-9d4b-4d62-be88-dd60993239ef\") " pod="openshift-marketplace/redhat-operators-g7mhn" Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.519346 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a8395da-9d4b-4d62-be88-dd60993239ef-utilities\") pod \"redhat-operators-g7mhn\" (UID: \"3a8395da-9d4b-4d62-be88-dd60993239ef\") " pod="openshift-marketplace/redhat-operators-g7mhn" Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.621288 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a8395da-9d4b-4d62-be88-dd60993239ef-utilities\") pod \"redhat-operators-g7mhn\" (UID: \"3a8395da-9d4b-4d62-be88-dd60993239ef\") " pod="openshift-marketplace/redhat-operators-g7mhn" Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.621375 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4vqt\" (UniqueName: \"kubernetes.io/projected/3a8395da-9d4b-4d62-be88-dd60993239ef-kube-api-access-l4vqt\") pod \"redhat-operators-g7mhn\" (UID: \"3a8395da-9d4b-4d62-be88-dd60993239ef\") " pod="openshift-marketplace/redhat-operators-g7mhn" Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.621398 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a8395da-9d4b-4d62-be88-dd60993239ef-catalog-content\") pod \"redhat-operators-g7mhn\" (UID: \"3a8395da-9d4b-4d62-be88-dd60993239ef\") " pod="openshift-marketplace/redhat-operators-g7mhn" Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.621894 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a8395da-9d4b-4d62-be88-dd60993239ef-utilities\") pod \"redhat-operators-g7mhn\" (UID: \"3a8395da-9d4b-4d62-be88-dd60993239ef\") " pod="openshift-marketplace/redhat-operators-g7mhn" Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.621942 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a8395da-9d4b-4d62-be88-dd60993239ef-catalog-content\") pod \"redhat-operators-g7mhn\" (UID: \"3a8395da-9d4b-4d62-be88-dd60993239ef\") " pod="openshift-marketplace/redhat-operators-g7mhn" Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.640746 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4vqt\" (UniqueName: \"kubernetes.io/projected/3a8395da-9d4b-4d62-be88-dd60993239ef-kube-api-access-l4vqt\") pod \"redhat-operators-g7mhn\" (UID: \"3a8395da-9d4b-4d62-be88-dd60993239ef\") " pod="openshift-marketplace/redhat-operators-g7mhn" Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.706111 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g7mhn" Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.982730 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t9xj4"] Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.984399 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t9xj4" Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.987287 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 10 19:01:36 crc kubenswrapper[4828]: I1210 19:01:36.989202 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t9xj4"] Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.086871 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g7mhn"] Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.130356 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca31d980-917d-4956-adb9-e31050c30f70-utilities\") pod \"community-operators-t9xj4\" (UID: \"ca31d980-917d-4956-adb9-e31050c30f70\") " pod="openshift-marketplace/community-operators-t9xj4" Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.130683 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcwf2\" (UniqueName: \"kubernetes.io/projected/ca31d980-917d-4956-adb9-e31050c30f70-kube-api-access-jcwf2\") pod \"community-operators-t9xj4\" (UID: \"ca31d980-917d-4956-adb9-e31050c30f70\") " pod="openshift-marketplace/community-operators-t9xj4" Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.130832 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca31d980-917d-4956-adb9-e31050c30f70-catalog-content\") pod \"community-operators-t9xj4\" (UID: \"ca31d980-917d-4956-adb9-e31050c30f70\") " pod="openshift-marketplace/community-operators-t9xj4" Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.231973 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcwf2\" (UniqueName: \"kubernetes.io/projected/ca31d980-917d-4956-adb9-e31050c30f70-kube-api-access-jcwf2\") pod \"community-operators-t9xj4\" (UID: \"ca31d980-917d-4956-adb9-e31050c30f70\") " pod="openshift-marketplace/community-operators-t9xj4" Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.232034 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca31d980-917d-4956-adb9-e31050c30f70-catalog-content\") pod \"community-operators-t9xj4\" (UID: \"ca31d980-917d-4956-adb9-e31050c30f70\") " pod="openshift-marketplace/community-operators-t9xj4" Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.232074 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca31d980-917d-4956-adb9-e31050c30f70-utilities\") pod \"community-operators-t9xj4\" (UID: \"ca31d980-917d-4956-adb9-e31050c30f70\") " pod="openshift-marketplace/community-operators-t9xj4" Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.232481 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca31d980-917d-4956-adb9-e31050c30f70-utilities\") pod \"community-operators-t9xj4\" (UID: \"ca31d980-917d-4956-adb9-e31050c30f70\") " pod="openshift-marketplace/community-operators-t9xj4" Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.232580 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca31d980-917d-4956-adb9-e31050c30f70-catalog-content\") pod \"community-operators-t9xj4\" (UID: \"ca31d980-917d-4956-adb9-e31050c30f70\") " pod="openshift-marketplace/community-operators-t9xj4" Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.251683 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcwf2\" (UniqueName: \"kubernetes.io/projected/ca31d980-917d-4956-adb9-e31050c30f70-kube-api-access-jcwf2\") pod \"community-operators-t9xj4\" (UID: \"ca31d980-917d-4956-adb9-e31050c30f70\") " pod="openshift-marketplace/community-operators-t9xj4" Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.339345 4828 generic.go:334] "Generic (PLEG): container finished" podID="3a8395da-9d4b-4d62-be88-dd60993239ef" containerID="43c98e13d168a86eda789b6db1e1fb8f7c8967f151f0c5aa30db6dfbeb9cba45" exitCode=0 Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.339388 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7mhn" event={"ID":"3a8395da-9d4b-4d62-be88-dd60993239ef","Type":"ContainerDied","Data":"43c98e13d168a86eda789b6db1e1fb8f7c8967f151f0c5aa30db6dfbeb9cba45"} Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.339428 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7mhn" event={"ID":"3a8395da-9d4b-4d62-be88-dd60993239ef","Type":"ContainerStarted","Data":"c209090aacdafb46c3c3fa0b623955b9559ae1e94efdaf3cd696b4b659fcd32e"} Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.342710 4828 generic.go:334] "Generic (PLEG): container finished" podID="b49ed428-86d9-4b1d-a2c6-f3f79fadb13d" containerID="837ad2895532b1937a4c09f08556137dac117a4db1eb23c5f2e6bcc7ed2552df" exitCode=0 Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.342785 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hm6hf" event={"ID":"b49ed428-86d9-4b1d-a2c6-f3f79fadb13d","Type":"ContainerDied","Data":"837ad2895532b1937a4c09f08556137dac117a4db1eb23c5f2e6bcc7ed2552df"} Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.348883 4828 generic.go:334] "Generic (PLEG): container finished" podID="fb55e51a-e3e7-4bea-8791-aec28f2b8873" containerID="6eb7e3703ece65e86d8492b10599d7ddc516421e03d563d18e3962c4ec3199a1" exitCode=0 Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.348915 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-df2d5" event={"ID":"fb55e51a-e3e7-4bea-8791-aec28f2b8873","Type":"ContainerDied","Data":"6eb7e3703ece65e86d8492b10599d7ddc516421e03d563d18e3962c4ec3199a1"} Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.407506 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t9xj4" Dec 10 19:01:37 crc kubenswrapper[4828]: I1210 19:01:37.785042 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t9xj4"] Dec 10 19:01:37 crc kubenswrapper[4828]: W1210 19:01:37.792114 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca31d980_917d_4956_adb9_e31050c30f70.slice/crio-fdb92dd9ae6b74800f771cf75036ed3b53ffbdd06167266384ff5df059c26296 WatchSource:0}: Error finding container fdb92dd9ae6b74800f771cf75036ed3b53ffbdd06167266384ff5df059c26296: Status 404 returned error can't find the container with id fdb92dd9ae6b74800f771cf75036ed3b53ffbdd06167266384ff5df059c26296 Dec 10 19:01:38 crc kubenswrapper[4828]: I1210 19:01:38.355411 4828 generic.go:334] "Generic (PLEG): container finished" podID="ca31d980-917d-4956-adb9-e31050c30f70" containerID="d30a2236386f3eea652cbec991865360404c42a197f093d140e3eb72c35ffede" exitCode=0 Dec 10 19:01:38 crc kubenswrapper[4828]: I1210 19:01:38.355506 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9xj4" event={"ID":"ca31d980-917d-4956-adb9-e31050c30f70","Type":"ContainerDied","Data":"d30a2236386f3eea652cbec991865360404c42a197f093d140e3eb72c35ffede"} Dec 10 19:01:38 crc kubenswrapper[4828]: I1210 19:01:38.357295 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9xj4" event={"ID":"ca31d980-917d-4956-adb9-e31050c30f70","Type":"ContainerStarted","Data":"fdb92dd9ae6b74800f771cf75036ed3b53ffbdd06167266384ff5df059c26296"} Dec 10 19:01:38 crc kubenswrapper[4828]: I1210 19:01:38.360182 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-df2d5" event={"ID":"fb55e51a-e3e7-4bea-8791-aec28f2b8873","Type":"ContainerStarted","Data":"6ff172691ea2d90f907319a20725476f03f42630c77617a061b307e6a1996464"} Dec 10 19:01:38 crc kubenswrapper[4828]: I1210 19:01:38.362642 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hm6hf" event={"ID":"b49ed428-86d9-4b1d-a2c6-f3f79fadb13d","Type":"ContainerStarted","Data":"93979656ee8125a4a64b266afcb4ea6356c889133bd1c243039c75f6537c991e"} Dec 10 19:01:38 crc kubenswrapper[4828]: I1210 19:01:38.394360 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-df2d5" podStartSLOduration=2.928678085 podStartE2EDuration="5.394341515s" podCreationTimestamp="2025-12-10 19:01:33 +0000 UTC" firstStartedPulling="2025-12-10 19:01:35.320397532 +0000 UTC m=+375.831008537" lastFinishedPulling="2025-12-10 19:01:37.786060962 +0000 UTC m=+378.296671967" observedRunningTime="2025-12-10 19:01:38.391504739 +0000 UTC m=+378.902115744" watchObservedRunningTime="2025-12-10 19:01:38.394341515 +0000 UTC m=+378.904952520" Dec 10 19:01:38 crc kubenswrapper[4828]: I1210 19:01:38.410968 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hm6hf" podStartSLOduration=2.5549939200000003 podStartE2EDuration="4.410948836s" podCreationTimestamp="2025-12-10 19:01:34 +0000 UTC" firstStartedPulling="2025-12-10 19:01:36.327768084 +0000 UTC m=+376.838379089" lastFinishedPulling="2025-12-10 19:01:38.183723 +0000 UTC m=+378.694334005" observedRunningTime="2025-12-10 19:01:38.407302276 +0000 UTC m=+378.917913281" watchObservedRunningTime="2025-12-10 19:01:38.410948836 +0000 UTC m=+378.921559841" Dec 10 19:01:39 crc kubenswrapper[4828]: I1210 19:01:39.369215 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7mhn" event={"ID":"3a8395da-9d4b-4d62-be88-dd60993239ef","Type":"ContainerStarted","Data":"8114cdcac4c1f2e74056f21cdf7c87a3ac9c5a9674e5a9d2598a81073a6010f8"} Dec 10 19:01:39 crc kubenswrapper[4828]: I1210 19:01:39.371018 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9xj4" event={"ID":"ca31d980-917d-4956-adb9-e31050c30f70","Type":"ContainerStarted","Data":"d02aca4882348efb8fd20c4b1ff0739438a23a79179e9039a8534db21422b5ab"} Dec 10 19:01:40 crc kubenswrapper[4828]: I1210 19:01:40.377119 4828 generic.go:334] "Generic (PLEG): container finished" podID="3a8395da-9d4b-4d62-be88-dd60993239ef" containerID="8114cdcac4c1f2e74056f21cdf7c87a3ac9c5a9674e5a9d2598a81073a6010f8" exitCode=0 Dec 10 19:01:40 crc kubenswrapper[4828]: I1210 19:01:40.377203 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7mhn" event={"ID":"3a8395da-9d4b-4d62-be88-dd60993239ef","Type":"ContainerDied","Data":"8114cdcac4c1f2e74056f21cdf7c87a3ac9c5a9674e5a9d2598a81073a6010f8"} Dec 10 19:01:40 crc kubenswrapper[4828]: I1210 19:01:40.379210 4828 generic.go:334] "Generic (PLEG): container finished" podID="ca31d980-917d-4956-adb9-e31050c30f70" containerID="d02aca4882348efb8fd20c4b1ff0739438a23a79179e9039a8534db21422b5ab" exitCode=0 Dec 10 19:01:40 crc kubenswrapper[4828]: I1210 19:01:40.379246 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9xj4" event={"ID":"ca31d980-917d-4956-adb9-e31050c30f70","Type":"ContainerDied","Data":"d02aca4882348efb8fd20c4b1ff0739438a23a79179e9039a8534db21422b5ab"} Dec 10 19:01:41 crc kubenswrapper[4828]: I1210 19:01:41.384770 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9xj4" event={"ID":"ca31d980-917d-4956-adb9-e31050c30f70","Type":"ContainerStarted","Data":"1dd5e36d05c02aeb28b6ad32d07ff32e7b04244f50e4013c36fe84ebd7ac3896"} Dec 10 19:01:41 crc kubenswrapper[4828]: I1210 19:01:41.386944 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7mhn" event={"ID":"3a8395da-9d4b-4d62-be88-dd60993239ef","Type":"ContainerStarted","Data":"350cfa1360e85b1bf666bba957451d1c0307b3ff6fde23a632bea378ad1eb706"} Dec 10 19:01:41 crc kubenswrapper[4828]: I1210 19:01:41.427607 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g7mhn" podStartSLOduration=1.953809584 podStartE2EDuration="5.42759051s" podCreationTimestamp="2025-12-10 19:01:36 +0000 UTC" firstStartedPulling="2025-12-10 19:01:37.341048426 +0000 UTC m=+377.851659431" lastFinishedPulling="2025-12-10 19:01:40.814829352 +0000 UTC m=+381.325440357" observedRunningTime="2025-12-10 19:01:41.427331232 +0000 UTC m=+381.937942237" watchObservedRunningTime="2025-12-10 19:01:41.42759051 +0000 UTC m=+381.938201515" Dec 10 19:01:41 crc kubenswrapper[4828]: I1210 19:01:41.428954 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t9xj4" podStartSLOduration=2.887489954 podStartE2EDuration="5.428947701s" podCreationTimestamp="2025-12-10 19:01:36 +0000 UTC" firstStartedPulling="2025-12-10 19:01:38.356755241 +0000 UTC m=+378.867366246" lastFinishedPulling="2025-12-10 19:01:40.898212998 +0000 UTC m=+381.408823993" observedRunningTime="2025-12-10 19:01:41.409190125 +0000 UTC m=+381.919801130" watchObservedRunningTime="2025-12-10 19:01:41.428947701 +0000 UTC m=+381.939558706" Dec 10 19:01:44 crc kubenswrapper[4828]: I1210 19:01:44.292574 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-df2d5" Dec 10 19:01:44 crc kubenswrapper[4828]: I1210 19:01:44.293320 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-df2d5" Dec 10 19:01:44 crc kubenswrapper[4828]: I1210 19:01:44.340099 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-df2d5" Dec 10 19:01:44 crc kubenswrapper[4828]: I1210 19:01:44.439577 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-df2d5" Dec 10 19:01:44 crc kubenswrapper[4828]: I1210 19:01:44.890230 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hm6hf" Dec 10 19:01:44 crc kubenswrapper[4828]: I1210 19:01:44.890302 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hm6hf" Dec 10 19:01:44 crc kubenswrapper[4828]: I1210 19:01:44.927944 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hm6hf" Dec 10 19:01:45 crc kubenswrapper[4828]: I1210 19:01:45.449116 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hm6hf" Dec 10 19:01:46 crc kubenswrapper[4828]: I1210 19:01:46.706898 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g7mhn" Dec 10 19:01:46 crc kubenswrapper[4828]: I1210 19:01:46.706944 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g7mhn" Dec 10 19:01:46 crc kubenswrapper[4828]: I1210 19:01:46.747762 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g7mhn" Dec 10 19:01:47 crc kubenswrapper[4828]: I1210 19:01:47.408288 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t9xj4" Dec 10 19:01:47 crc kubenswrapper[4828]: I1210 19:01:47.408614 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t9xj4" Dec 10 19:01:47 crc kubenswrapper[4828]: I1210 19:01:47.450269 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g7mhn" Dec 10 19:01:47 crc kubenswrapper[4828]: I1210 19:01:47.450376 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t9xj4" Dec 10 19:01:47 crc kubenswrapper[4828]: I1210 19:01:47.489700 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t9xj4" Dec 10 19:01:48 crc kubenswrapper[4828]: I1210 19:01:48.016375 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" podUID="1e72c425-11b5-4fd4-9f71-9df4d4489475" containerName="registry" containerID="cri-o://be827b35061551e27e2ea93f8ae82720cf53c1222786b1adcdb55d1302abee78" gracePeriod=30 Dec 10 19:01:49 crc kubenswrapper[4828]: I1210 19:01:49.426473 4828 generic.go:334] "Generic (PLEG): container finished" podID="1e72c425-11b5-4fd4-9f71-9df4d4489475" containerID="be827b35061551e27e2ea93f8ae82720cf53c1222786b1adcdb55d1302abee78" exitCode=0 Dec 10 19:01:49 crc kubenswrapper[4828]: I1210 19:01:49.426517 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" event={"ID":"1e72c425-11b5-4fd4-9f71-9df4d4489475","Type":"ContainerDied","Data":"be827b35061551e27e2ea93f8ae82720cf53c1222786b1adcdb55d1302abee78"} Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.432912 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" event={"ID":"1e72c425-11b5-4fd4-9f71-9df4d4489475","Type":"ContainerDied","Data":"2db450e3fd742a7c15741e18ba9365b0f4caa86e1cbcb37f123da33ddd3de90d"} Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.433241 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2db450e3fd742a7c15741e18ba9365b0f4caa86e1cbcb37f123da33ddd3de90d" Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.450050 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.610644 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1e72c425-11b5-4fd4-9f71-9df4d4489475-installation-pull-secrets\") pod \"1e72c425-11b5-4fd4-9f71-9df4d4489475\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.610714 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1e72c425-11b5-4fd4-9f71-9df4d4489475-trusted-ca\") pod \"1e72c425-11b5-4fd4-9f71-9df4d4489475\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.610744 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmz86\" (UniqueName: \"kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-kube-api-access-cmz86\") pod \"1e72c425-11b5-4fd4-9f71-9df4d4489475\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.610811 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-registry-tls\") pod \"1e72c425-11b5-4fd4-9f71-9df4d4489475\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.610872 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1e72c425-11b5-4fd4-9f71-9df4d4489475-registry-certificates\") pod \"1e72c425-11b5-4fd4-9f71-9df4d4489475\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.610898 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-bound-sa-token\") pod \"1e72c425-11b5-4fd4-9f71-9df4d4489475\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.611100 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"1e72c425-11b5-4fd4-9f71-9df4d4489475\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.611132 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1e72c425-11b5-4fd4-9f71-9df4d4489475-ca-trust-extracted\") pod \"1e72c425-11b5-4fd4-9f71-9df4d4489475\" (UID: \"1e72c425-11b5-4fd4-9f71-9df4d4489475\") " Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.611619 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e72c425-11b5-4fd4-9f71-9df4d4489475-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "1e72c425-11b5-4fd4-9f71-9df4d4489475" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.611758 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e72c425-11b5-4fd4-9f71-9df4d4489475-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "1e72c425-11b5-4fd4-9f71-9df4d4489475" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.619023 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "1e72c425-11b5-4fd4-9f71-9df4d4489475" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.620028 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-kube-api-access-cmz86" (OuterVolumeSpecName: "kube-api-access-cmz86") pod "1e72c425-11b5-4fd4-9f71-9df4d4489475" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475"). InnerVolumeSpecName "kube-api-access-cmz86". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.620545 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e72c425-11b5-4fd4-9f71-9df4d4489475-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "1e72c425-11b5-4fd4-9f71-9df4d4489475" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.625776 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "1e72c425-11b5-4fd4-9f71-9df4d4489475" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.635288 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e72c425-11b5-4fd4-9f71-9df4d4489475-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "1e72c425-11b5-4fd4-9f71-9df4d4489475" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.653661 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "1e72c425-11b5-4fd4-9f71-9df4d4489475" (UID: "1e72c425-11b5-4fd4-9f71-9df4d4489475"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.712056 4828 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.712093 4828 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1e72c425-11b5-4fd4-9f71-9df4d4489475-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.712103 4828 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1e72c425-11b5-4fd4-9f71-9df4d4489475-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.712113 4828 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1e72c425-11b5-4fd4-9f71-9df4d4489475-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.712122 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmz86\" (UniqueName: \"kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-kube-api-access-cmz86\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.712130 4828 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1e72c425-11b5-4fd4-9f71-9df4d4489475-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:50 crc kubenswrapper[4828]: I1210 19:01:50.712138 4828 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1e72c425-11b5-4fd4-9f71-9df4d4489475-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 10 19:01:51 crc kubenswrapper[4828]: I1210 19:01:51.231058 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:01:51 crc kubenswrapper[4828]: I1210 19:01:51.231175 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:01:51 crc kubenswrapper[4828]: I1210 19:01:51.439376 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-c4wpt" Dec 10 19:01:51 crc kubenswrapper[4828]: I1210 19:01:51.484412 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c4wpt"] Dec 10 19:01:51 crc kubenswrapper[4828]: I1210 19:01:51.486090 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-c4wpt"] Dec 10 19:01:51 crc kubenswrapper[4828]: I1210 19:01:51.794662 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e72c425-11b5-4fd4-9f71-9df4d4489475" path="/var/lib/kubelet/pods/1e72c425-11b5-4fd4-9f71-9df4d4489475/volumes" Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.763090 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-fhkjd"] Dec 10 19:02:01 crc kubenswrapper[4828]: E1210 19:02:01.763951 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e72c425-11b5-4fd4-9f71-9df4d4489475" containerName="registry" Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.763970 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e72c425-11b5-4fd4-9f71-9df4d4489475" containerName="registry" Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.764084 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e72c425-11b5-4fd4-9f71-9df4d4489475" containerName="registry" Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.764475 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-fhkjd" Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.766687 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemetry-config" Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.766927 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-root-ca.crt" Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.769722 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-dockercfg-wwt9l" Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.769739 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-fhkjd"] Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.770139 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"openshift-service-ca.crt" Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.770406 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-tls" Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.870870 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zntgp\" (UniqueName: \"kubernetes.io/projected/365dd60f-b5fd-4532-96d8-00bb9aec49ba-kube-api-access-zntgp\") pod \"cluster-monitoring-operator-6d5b84845-fhkjd\" (UID: \"365dd60f-b5fd-4532-96d8-00bb9aec49ba\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-fhkjd" Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.870979 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/365dd60f-b5fd-4532-96d8-00bb9aec49ba-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-fhkjd\" (UID: \"365dd60f-b5fd-4532-96d8-00bb9aec49ba\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-fhkjd" Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.871030 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/365dd60f-b5fd-4532-96d8-00bb9aec49ba-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-fhkjd\" (UID: \"365dd60f-b5fd-4532-96d8-00bb9aec49ba\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-fhkjd" Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.972097 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zntgp\" (UniqueName: \"kubernetes.io/projected/365dd60f-b5fd-4532-96d8-00bb9aec49ba-kube-api-access-zntgp\") pod \"cluster-monitoring-operator-6d5b84845-fhkjd\" (UID: \"365dd60f-b5fd-4532-96d8-00bb9aec49ba\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-fhkjd" Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.972190 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/365dd60f-b5fd-4532-96d8-00bb9aec49ba-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-fhkjd\" (UID: \"365dd60f-b5fd-4532-96d8-00bb9aec49ba\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-fhkjd" Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.972276 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/365dd60f-b5fd-4532-96d8-00bb9aec49ba-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-fhkjd\" (UID: \"365dd60f-b5fd-4532-96d8-00bb9aec49ba\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-fhkjd" Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.974208 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/365dd60f-b5fd-4532-96d8-00bb9aec49ba-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-fhkjd\" (UID: \"365dd60f-b5fd-4532-96d8-00bb9aec49ba\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-fhkjd" Dec 10 19:02:01 crc kubenswrapper[4828]: I1210 19:02:01.981389 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/365dd60f-b5fd-4532-96d8-00bb9aec49ba-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-fhkjd\" (UID: \"365dd60f-b5fd-4532-96d8-00bb9aec49ba\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-fhkjd" Dec 10 19:02:02 crc kubenswrapper[4828]: I1210 19:02:02.006261 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zntgp\" (UniqueName: \"kubernetes.io/projected/365dd60f-b5fd-4532-96d8-00bb9aec49ba-kube-api-access-zntgp\") pod \"cluster-monitoring-operator-6d5b84845-fhkjd\" (UID: \"365dd60f-b5fd-4532-96d8-00bb9aec49ba\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-fhkjd" Dec 10 19:02:02 crc kubenswrapper[4828]: I1210 19:02:02.085325 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-fhkjd" Dec 10 19:02:02 crc kubenswrapper[4828]: I1210 19:02:02.270710 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-fhkjd"] Dec 10 19:02:02 crc kubenswrapper[4828]: W1210 19:02:02.275719 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod365dd60f_b5fd_4532_96d8_00bb9aec49ba.slice/crio-2423b99d9d2fe186b1ea50d33508770a446b869c24ff868bba07fbdc09ed35d5 WatchSource:0}: Error finding container 2423b99d9d2fe186b1ea50d33508770a446b869c24ff868bba07fbdc09ed35d5: Status 404 returned error can't find the container with id 2423b99d9d2fe186b1ea50d33508770a446b869c24ff868bba07fbdc09ed35d5 Dec 10 19:02:02 crc kubenswrapper[4828]: I1210 19:02:02.500710 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-fhkjd" event={"ID":"365dd60f-b5fd-4532-96d8-00bb9aec49ba","Type":"ContainerStarted","Data":"2423b99d9d2fe186b1ea50d33508770a446b869c24ff868bba07fbdc09ed35d5"} Dec 10 19:02:04 crc kubenswrapper[4828]: I1210 19:02:04.511604 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-fhkjd" event={"ID":"365dd60f-b5fd-4532-96d8-00bb9aec49ba","Type":"ContainerStarted","Data":"22287b51c00284cb6b7e3f58e05e417ce1cec8ad51d23c7f3655a9cf5ff34d47"} Dec 10 19:02:04 crc kubenswrapper[4828]: I1210 19:02:04.525967 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-fhkjd" podStartSLOduration=1.714783052 podStartE2EDuration="3.525948333s" podCreationTimestamp="2025-12-10 19:02:01 +0000 UTC" firstStartedPulling="2025-12-10 19:02:02.278245138 +0000 UTC m=+402.788856153" lastFinishedPulling="2025-12-10 19:02:04.089410419 +0000 UTC m=+404.600021434" observedRunningTime="2025-12-10 19:02:04.524914632 +0000 UTC m=+405.035525637" watchObservedRunningTime="2025-12-10 19:02:04.525948333 +0000 UTC m=+405.036559338" Dec 10 19:02:04 crc kubenswrapper[4828]: I1210 19:02:04.655826 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-prp6r"] Dec 10 19:02:04 crc kubenswrapper[4828]: I1210 19:02:04.656435 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-prp6r" Dec 10 19:02:04 crc kubenswrapper[4828]: I1210 19:02:04.658018 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-tls" Dec 10 19:02:04 crc kubenswrapper[4828]: I1210 19:02:04.658673 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-dockercfg-pgz7g" Dec 10 19:02:04 crc kubenswrapper[4828]: I1210 19:02:04.662875 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-prp6r"] Dec 10 19:02:04 crc kubenswrapper[4828]: I1210 19:02:04.807304 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/4ba52117-fd7d-4274-86a3-b443ff3fb2a8-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-prp6r\" (UID: \"4ba52117-fd7d-4274-86a3-b443ff3fb2a8\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-prp6r" Dec 10 19:02:04 crc kubenswrapper[4828]: I1210 19:02:04.911237 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/4ba52117-fd7d-4274-86a3-b443ff3fb2a8-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-prp6r\" (UID: \"4ba52117-fd7d-4274-86a3-b443ff3fb2a8\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-prp6r" Dec 10 19:02:04 crc kubenswrapper[4828]: E1210 19:02:04.912200 4828 secret.go:188] Couldn't get secret openshift-monitoring/prometheus-operator-admission-webhook-tls: secret "prometheus-operator-admission-webhook-tls" not found Dec 10 19:02:04 crc kubenswrapper[4828]: E1210 19:02:04.912253 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4ba52117-fd7d-4274-86a3-b443ff3fb2a8-tls-certificates podName:4ba52117-fd7d-4274-86a3-b443ff3fb2a8 nodeName:}" failed. No retries permitted until 2025-12-10 19:02:05.4122379 +0000 UTC m=+405.922848905 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-certificates" (UniqueName: "kubernetes.io/secret/4ba52117-fd7d-4274-86a3-b443ff3fb2a8-tls-certificates") pod "prometheus-operator-admission-webhook-f54c54754-prp6r" (UID: "4ba52117-fd7d-4274-86a3-b443ff3fb2a8") : secret "prometheus-operator-admission-webhook-tls" not found Dec 10 19:02:05 crc kubenswrapper[4828]: I1210 19:02:05.416230 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/4ba52117-fd7d-4274-86a3-b443ff3fb2a8-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-prp6r\" (UID: \"4ba52117-fd7d-4274-86a3-b443ff3fb2a8\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-prp6r" Dec 10 19:02:05 crc kubenswrapper[4828]: I1210 19:02:05.422974 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/4ba52117-fd7d-4274-86a3-b443ff3fb2a8-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-prp6r\" (UID: \"4ba52117-fd7d-4274-86a3-b443ff3fb2a8\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-prp6r" Dec 10 19:02:05 crc kubenswrapper[4828]: I1210 19:02:05.570734 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-prp6r" Dec 10 19:02:05 crc kubenswrapper[4828]: I1210 19:02:05.745726 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-prp6r"] Dec 10 19:02:05 crc kubenswrapper[4828]: W1210 19:02:05.750658 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ba52117_fd7d_4274_86a3_b443ff3fb2a8.slice/crio-397ff09dc8c5008b406a65743e8385a56abb30c3566cf9568d5a1dc280ded088 WatchSource:0}: Error finding container 397ff09dc8c5008b406a65743e8385a56abb30c3566cf9568d5a1dc280ded088: Status 404 returned error can't find the container with id 397ff09dc8c5008b406a65743e8385a56abb30c3566cf9568d5a1dc280ded088 Dec 10 19:02:06 crc kubenswrapper[4828]: I1210 19:02:06.524312 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-prp6r" event={"ID":"4ba52117-fd7d-4274-86a3-b443ff3fb2a8","Type":"ContainerStarted","Data":"397ff09dc8c5008b406a65743e8385a56abb30c3566cf9568d5a1dc280ded088"} Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.545766 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-prp6r" event={"ID":"4ba52117-fd7d-4274-86a3-b443ff3fb2a8","Type":"ContainerStarted","Data":"0eaaa4b98c902b3b5b50da85d405b737077e1820341e57957b237270ab3c2c06"} Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.546433 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-prp6r" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.551536 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-prp6r" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.571247 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-prp6r" podStartSLOduration=2.199430545 podStartE2EDuration="5.57122337s" podCreationTimestamp="2025-12-10 19:02:04 +0000 UTC" firstStartedPulling="2025-12-10 19:02:05.752635629 +0000 UTC m=+406.263246634" lastFinishedPulling="2025-12-10 19:02:09.124428454 +0000 UTC m=+409.635039459" observedRunningTime="2025-12-10 19:02:09.567662766 +0000 UTC m=+410.078273771" watchObservedRunningTime="2025-12-10 19:02:09.57122337 +0000 UTC m=+410.081834385" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.705759 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-j9mx7"] Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.707409 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.712458 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-tls" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.712495 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-dockercfg-n6b9p" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.712737 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-kube-rbac-proxy-config" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.712742 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-client-ca" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.724039 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-j9mx7"] Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.767670 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/9f9e4fab-4e74-4968-95fa-fa0f512fcfb1-metrics-client-ca\") pod \"prometheus-operator-db54df47d-j9mx7\" (UID: \"9f9e4fab-4e74-4968-95fa-fa0f512fcfb1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.767721 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9f9e4fab-4e74-4968-95fa-fa0f512fcfb1-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-j9mx7\" (UID: \"9f9e4fab-4e74-4968-95fa-fa0f512fcfb1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.767929 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f9e4fab-4e74-4968-95fa-fa0f512fcfb1-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-j9mx7\" (UID: \"9f9e4fab-4e74-4968-95fa-fa0f512fcfb1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.768028 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cdfd\" (UniqueName: \"kubernetes.io/projected/9f9e4fab-4e74-4968-95fa-fa0f512fcfb1-kube-api-access-2cdfd\") pod \"prometheus-operator-db54df47d-j9mx7\" (UID: \"9f9e4fab-4e74-4968-95fa-fa0f512fcfb1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.869266 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cdfd\" (UniqueName: \"kubernetes.io/projected/9f9e4fab-4e74-4968-95fa-fa0f512fcfb1-kube-api-access-2cdfd\") pod \"prometheus-operator-db54df47d-j9mx7\" (UID: \"9f9e4fab-4e74-4968-95fa-fa0f512fcfb1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.869323 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/9f9e4fab-4e74-4968-95fa-fa0f512fcfb1-metrics-client-ca\") pod \"prometheus-operator-db54df47d-j9mx7\" (UID: \"9f9e4fab-4e74-4968-95fa-fa0f512fcfb1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.869346 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9f9e4fab-4e74-4968-95fa-fa0f512fcfb1-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-j9mx7\" (UID: \"9f9e4fab-4e74-4968-95fa-fa0f512fcfb1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.869408 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f9e4fab-4e74-4968-95fa-fa0f512fcfb1-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-j9mx7\" (UID: \"9f9e4fab-4e74-4968-95fa-fa0f512fcfb1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.870329 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/9f9e4fab-4e74-4968-95fa-fa0f512fcfb1-metrics-client-ca\") pod \"prometheus-operator-db54df47d-j9mx7\" (UID: \"9f9e4fab-4e74-4968-95fa-fa0f512fcfb1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.875145 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f9e4fab-4e74-4968-95fa-fa0f512fcfb1-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-j9mx7\" (UID: \"9f9e4fab-4e74-4968-95fa-fa0f512fcfb1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.875145 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9f9e4fab-4e74-4968-95fa-fa0f512fcfb1-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-j9mx7\" (UID: \"9f9e4fab-4e74-4968-95fa-fa0f512fcfb1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" Dec 10 19:02:09 crc kubenswrapper[4828]: I1210 19:02:09.886211 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cdfd\" (UniqueName: \"kubernetes.io/projected/9f9e4fab-4e74-4968-95fa-fa0f512fcfb1-kube-api-access-2cdfd\") pod \"prometheus-operator-db54df47d-j9mx7\" (UID: \"9f9e4fab-4e74-4968-95fa-fa0f512fcfb1\") " pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" Dec 10 19:02:10 crc kubenswrapper[4828]: I1210 19:02:10.030293 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" Dec 10 19:02:10 crc kubenswrapper[4828]: I1210 19:02:10.213958 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-j9mx7"] Dec 10 19:02:10 crc kubenswrapper[4828]: W1210 19:02:10.219195 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f9e4fab_4e74_4968_95fa_fa0f512fcfb1.slice/crio-b7b4aa970498e754a1a14ed115bf79eb36cf95d74088d63ec97e79e178bdb412 WatchSource:0}: Error finding container b7b4aa970498e754a1a14ed115bf79eb36cf95d74088d63ec97e79e178bdb412: Status 404 returned error can't find the container with id b7b4aa970498e754a1a14ed115bf79eb36cf95d74088d63ec97e79e178bdb412 Dec 10 19:02:11 crc kubenswrapper[4828]: I1210 19:02:10.557592 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" event={"ID":"9f9e4fab-4e74-4968-95fa-fa0f512fcfb1","Type":"ContainerStarted","Data":"b7b4aa970498e754a1a14ed115bf79eb36cf95d74088d63ec97e79e178bdb412"} Dec 10 19:02:15 crc kubenswrapper[4828]: I1210 19:02:15.590837 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" event={"ID":"9f9e4fab-4e74-4968-95fa-fa0f512fcfb1","Type":"ContainerStarted","Data":"7d51fe33310ea23f8a48382931113e44ef473b112d1b09c3b0c1a3335868e146"} Dec 10 19:02:15 crc kubenswrapper[4828]: I1210 19:02:15.591732 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" event={"ID":"9f9e4fab-4e74-4968-95fa-fa0f512fcfb1","Type":"ContainerStarted","Data":"d8ac583fd610c9e15a6cb27029df08bb48fa7c62a74296768edf28bdbf5b2bf3"} Dec 10 19:02:15 crc kubenswrapper[4828]: I1210 19:02:15.614710 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-db54df47d-j9mx7" podStartSLOduration=2.215539555 podStartE2EDuration="6.614691626s" podCreationTimestamp="2025-12-10 19:02:09 +0000 UTC" firstStartedPulling="2025-12-10 19:02:10.220468849 +0000 UTC m=+410.731079844" lastFinishedPulling="2025-12-10 19:02:14.6196209 +0000 UTC m=+415.130231915" observedRunningTime="2025-12-10 19:02:15.612341887 +0000 UTC m=+416.122952902" watchObservedRunningTime="2025-12-10 19:02:15.614691626 +0000 UTC m=+416.125302641" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.067716 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv"] Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.069011 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.070709 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-kube-rbac-proxy-config" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.071487 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-tls" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.072214 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-dockercfg-nfbfs" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.082400 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w"] Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.083358 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.085439 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-dockercfg-tpqdn" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.085626 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-kube-rbac-proxy-config" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.085835 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-state-metrics-custom-resource-state-configmap" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.085961 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-tls" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.089868 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv"] Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.109792 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-nxjxg"] Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.111408 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.114069 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-tls" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.114089 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-dockercfg-ghgvk" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.114113 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-kube-rbac-proxy-config" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.121360 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w"] Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.263870 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/f47fe344-778e-4d41-aede-dd74a4c8d1a4-node-exporter-wtmp\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.263925 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/f47fe344-778e-4d41-aede-dd74a4c8d1a4-node-exporter-textfile\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.263964 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b067e07f-eb76-4de2-a145-e5a4d0770183-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.264007 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f47fe344-778e-4d41-aede-dd74a4c8d1a4-metrics-client-ca\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.264033 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/f47fe344-778e-4d41-aede-dd74a4c8d1a4-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.264169 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/f47fe344-778e-4d41-aede-dd74a4c8d1a4-node-exporter-tls\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.264228 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/9841286a-46d2-4d2c-be70-834b9bcda856-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-wwqgv\" (UID: \"9841286a-46d2-4d2c-be70-834b9bcda856\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.264271 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2wt2\" (UniqueName: \"kubernetes.io/projected/b067e07f-eb76-4de2-a145-e5a4d0770183-kube-api-access-c2wt2\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.264305 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/b067e07f-eb76-4de2-a145-e5a4d0770183-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.264347 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c92pt\" (UniqueName: \"kubernetes.io/projected/f47fe344-778e-4d41-aede-dd74a4c8d1a4-kube-api-access-c92pt\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.264418 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pj82\" (UniqueName: \"kubernetes.io/projected/9841286a-46d2-4d2c-be70-834b9bcda856-kube-api-access-9pj82\") pod \"openshift-state-metrics-566fddb674-wwqgv\" (UID: \"9841286a-46d2-4d2c-be70-834b9bcda856\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.264475 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/b067e07f-eb76-4de2-a145-e5a4d0770183-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.264556 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/f47fe344-778e-4d41-aede-dd74a4c8d1a4-root\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.264582 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9841286a-46d2-4d2c-be70-834b9bcda856-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-wwqgv\" (UID: \"9841286a-46d2-4d2c-be70-834b9bcda856\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.264616 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/b067e07f-eb76-4de2-a145-e5a4d0770183-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.264636 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b067e07f-eb76-4de2-a145-e5a4d0770183-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.264655 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9841286a-46d2-4d2c-be70-834b9bcda856-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-wwqgv\" (UID: \"9841286a-46d2-4d2c-be70-834b9bcda856\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.264676 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f47fe344-778e-4d41-aede-dd74a4c8d1a4-sys\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.365662 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/b067e07f-eb76-4de2-a145-e5a4d0770183-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.365749 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/f47fe344-778e-4d41-aede-dd74a4c8d1a4-root\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.365778 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9841286a-46d2-4d2c-be70-834b9bcda856-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-wwqgv\" (UID: \"9841286a-46d2-4d2c-be70-834b9bcda856\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.365842 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/b067e07f-eb76-4de2-a145-e5a4d0770183-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.365865 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b067e07f-eb76-4de2-a145-e5a4d0770183-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.365890 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9841286a-46d2-4d2c-be70-834b9bcda856-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-wwqgv\" (UID: \"9841286a-46d2-4d2c-be70-834b9bcda856\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.365894 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/f47fe344-778e-4d41-aede-dd74a4c8d1a4-root\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.365915 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f47fe344-778e-4d41-aede-dd74a4c8d1a4-sys\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.365947 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f47fe344-778e-4d41-aede-dd74a4c8d1a4-sys\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.366006 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/f47fe344-778e-4d41-aede-dd74a4c8d1a4-node-exporter-wtmp\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.366033 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/f47fe344-778e-4d41-aede-dd74a4c8d1a4-node-exporter-textfile\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.366071 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b067e07f-eb76-4de2-a145-e5a4d0770183-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.366089 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f47fe344-778e-4d41-aede-dd74a4c8d1a4-metrics-client-ca\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.366110 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/f47fe344-778e-4d41-aede-dd74a4c8d1a4-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.366152 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/f47fe344-778e-4d41-aede-dd74a4c8d1a4-node-exporter-tls\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.366168 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/9841286a-46d2-4d2c-be70-834b9bcda856-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-wwqgv\" (UID: \"9841286a-46d2-4d2c-be70-834b9bcda856\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.366192 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2wt2\" (UniqueName: \"kubernetes.io/projected/b067e07f-eb76-4de2-a145-e5a4d0770183-kube-api-access-c2wt2\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.366215 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/b067e07f-eb76-4de2-a145-e5a4d0770183-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.366237 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pj82\" (UniqueName: \"kubernetes.io/projected/9841286a-46d2-4d2c-be70-834b9bcda856-kube-api-access-9pj82\") pod \"openshift-state-metrics-566fddb674-wwqgv\" (UID: \"9841286a-46d2-4d2c-be70-834b9bcda856\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.366253 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c92pt\" (UniqueName: \"kubernetes.io/projected/f47fe344-778e-4d41-aede-dd74a4c8d1a4-kube-api-access-c92pt\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.366290 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/f47fe344-778e-4d41-aede-dd74a4c8d1a4-node-exporter-wtmp\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: E1210 19:02:17.366379 4828 secret.go:188] Couldn't get secret openshift-monitoring/node-exporter-tls: secret "node-exporter-tls" not found Dec 10 19:02:17 crc kubenswrapper[4828]: E1210 19:02:17.366438 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f47fe344-778e-4d41-aede-dd74a4c8d1a4-node-exporter-tls podName:f47fe344-778e-4d41-aede-dd74a4c8d1a4 nodeName:}" failed. No retries permitted until 2025-12-10 19:02:17.866423099 +0000 UTC m=+418.377034104 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-exporter-tls" (UniqueName: "kubernetes.io/secret/f47fe344-778e-4d41-aede-dd74a4c8d1a4-node-exporter-tls") pod "node-exporter-nxjxg" (UID: "f47fe344-778e-4d41-aede-dd74a4c8d1a4") : secret "node-exporter-tls" not found Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.366448 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/b067e07f-eb76-4de2-a145-e5a4d0770183-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.366822 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b067e07f-eb76-4de2-a145-e5a4d0770183-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.366834 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/b067e07f-eb76-4de2-a145-e5a4d0770183-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.367135 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f47fe344-778e-4d41-aede-dd74a4c8d1a4-metrics-client-ca\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.367140 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/f47fe344-778e-4d41-aede-dd74a4c8d1a4-node-exporter-textfile\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.367215 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/9841286a-46d2-4d2c-be70-834b9bcda856-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-wwqgv\" (UID: \"9841286a-46d2-4d2c-be70-834b9bcda856\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.372297 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9841286a-46d2-4d2c-be70-834b9bcda856-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-wwqgv\" (UID: \"9841286a-46d2-4d2c-be70-834b9bcda856\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.373664 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9841286a-46d2-4d2c-be70-834b9bcda856-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-wwqgv\" (UID: \"9841286a-46d2-4d2c-be70-834b9bcda856\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.375143 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b067e07f-eb76-4de2-a145-e5a4d0770183-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.376105 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/f47fe344-778e-4d41-aede-dd74a4c8d1a4-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.377730 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/b067e07f-eb76-4de2-a145-e5a4d0770183-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.385552 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2wt2\" (UniqueName: \"kubernetes.io/projected/b067e07f-eb76-4de2-a145-e5a4d0770183-kube-api-access-c2wt2\") pod \"kube-state-metrics-777cb5bd5d-snw8w\" (UID: \"b067e07f-eb76-4de2-a145-e5a4d0770183\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.387164 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c92pt\" (UniqueName: \"kubernetes.io/projected/f47fe344-778e-4d41-aede-dd74a4c8d1a4-kube-api-access-c92pt\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.391484 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pj82\" (UniqueName: \"kubernetes.io/projected/9841286a-46d2-4d2c-be70-834b9bcda856-kube-api-access-9pj82\") pod \"openshift-state-metrics-566fddb674-wwqgv\" (UID: \"9841286a-46d2-4d2c-be70-834b9bcda856\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.397884 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.685814 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.821813 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w"] Dec 10 19:02:17 crc kubenswrapper[4828]: W1210 19:02:17.839908 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb067e07f_eb76_4de2_a145_e5a4d0770183.slice/crio-8ed4af2db63577128f6e1ea4b495e30dd14aa7379046fd821cf1c7e913e48d10 WatchSource:0}: Error finding container 8ed4af2db63577128f6e1ea4b495e30dd14aa7379046fd821cf1c7e913e48d10: Status 404 returned error can't find the container with id 8ed4af2db63577128f6e1ea4b495e30dd14aa7379046fd821cf1c7e913e48d10 Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.873885 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/f47fe344-778e-4d41-aede-dd74a4c8d1a4-node-exporter-tls\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.880706 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/f47fe344-778e-4d41-aede-dd74a4c8d1a4-node-exporter-tls\") pod \"node-exporter-nxjxg\" (UID: \"f47fe344-778e-4d41-aede-dd74a4c8d1a4\") " pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:17 crc kubenswrapper[4828]: I1210 19:02:17.905503 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv"] Dec 10 19:02:17 crc kubenswrapper[4828]: W1210 19:02:17.912849 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9841286a_46d2_4d2c_be70_834b9bcda856.slice/crio-8df9dcc4e8870fd10158f982e74661a19a58f35d77ec80f7bd05c646088fa0b8 WatchSource:0}: Error finding container 8df9dcc4e8870fd10158f982e74661a19a58f35d77ec80f7bd05c646088fa0b8: Status 404 returned error can't find the container with id 8df9dcc4e8870fd10158f982e74661a19a58f35d77ec80f7bd05c646088fa0b8 Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.026416 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-nxjxg" Dec 10 19:02:18 crc kubenswrapper[4828]: W1210 19:02:18.043129 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf47fe344_778e_4d41_aede_dd74a4c8d1a4.slice/crio-af567e56de673fa1ac753ffaeb0fe9abc5b49ae154446fab806d38df1aa1ad96 WatchSource:0}: Error finding container af567e56de673fa1ac753ffaeb0fe9abc5b49ae154446fab806d38df1aa1ad96: Status 404 returned error can't find the container with id af567e56de673fa1ac753ffaeb0fe9abc5b49ae154446fab806d38df1aa1ad96 Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.127059 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.130786 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.133390 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-generated" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.133671 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-web" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.133986 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.135147 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-metric" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.135515 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-web-config" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.136410 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-dockercfg-lptxv" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.136577 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.137687 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls-assets-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.140842 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"alertmanager-trusted-ca-bundle" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.151661 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.278056 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-web-config\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.278107 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kknnw\" (UniqueName: \"kubernetes.io/projected/f8ab78da-de77-4dd1-9fd0-822314f0691e-kube-api-access-kknnw\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.278161 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f8ab78da-de77-4dd1-9fd0-822314f0691e-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.278199 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-config-volume\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.278219 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/f8ab78da-de77-4dd1-9fd0-822314f0691e-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.278250 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.278288 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f8ab78da-de77-4dd1-9fd0-822314f0691e-tls-assets\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.278314 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.278335 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.278360 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f8ab78da-de77-4dd1-9fd0-822314f0691e-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.278412 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f8ab78da-de77-4dd1-9fd0-822314f0691e-config-out\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.278479 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.379925 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f8ab78da-de77-4dd1-9fd0-822314f0691e-tls-assets\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.379977 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.380002 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.380021 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f8ab78da-de77-4dd1-9fd0-822314f0691e-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.380039 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f8ab78da-de77-4dd1-9fd0-822314f0691e-config-out\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.380069 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.380095 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-web-config\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.380113 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kknnw\" (UniqueName: \"kubernetes.io/projected/f8ab78da-de77-4dd1-9fd0-822314f0691e-kube-api-access-kknnw\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.380150 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f8ab78da-de77-4dd1-9fd0-822314f0691e-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.380164 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-config-volume\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.380180 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/f8ab78da-de77-4dd1-9fd0-822314f0691e-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.380199 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.381311 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f8ab78da-de77-4dd1-9fd0-822314f0691e-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.381605 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/f8ab78da-de77-4dd1-9fd0-822314f0691e-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.381656 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/f8ab78da-de77-4dd1-9fd0-822314f0691e-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.385025 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-web-config\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.385071 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.385127 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.385450 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/f8ab78da-de77-4dd1-9fd0-822314f0691e-config-out\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.385590 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.386204 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/f8ab78da-de77-4dd1-9fd0-822314f0691e-tls-assets\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.387918 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-config-volume\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.388443 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/f8ab78da-de77-4dd1-9fd0-822314f0691e-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.407779 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kknnw\" (UniqueName: \"kubernetes.io/projected/f8ab78da-de77-4dd1-9fd0-822314f0691e-kube-api-access-kknnw\") pod \"alertmanager-main-0\" (UID: \"f8ab78da-de77-4dd1-9fd0-822314f0691e\") " pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.449916 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.619639 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-nxjxg" event={"ID":"f47fe344-778e-4d41-aede-dd74a4c8d1a4","Type":"ContainerStarted","Data":"af567e56de673fa1ac753ffaeb0fe9abc5b49ae154446fab806d38df1aa1ad96"} Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.622049 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" event={"ID":"9841286a-46d2-4d2c-be70-834b9bcda856","Type":"ContainerStarted","Data":"1a774468cc335219b7f8517136d61cc237b88c27bc571b459909683f33e313a8"} Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.622108 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" event={"ID":"9841286a-46d2-4d2c-be70-834b9bcda856","Type":"ContainerStarted","Data":"9317f783ac5f3fef445ec87bc85f34ed8caf5b553e84921ffaa37f6ecc1cb28d"} Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.622125 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" event={"ID":"9841286a-46d2-4d2c-be70-834b9bcda856","Type":"ContainerStarted","Data":"8df9dcc4e8870fd10158f982e74661a19a58f35d77ec80f7bd05c646088fa0b8"} Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.623872 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" event={"ID":"b067e07f-eb76-4de2-a145-e5a4d0770183","Type":"ContainerStarted","Data":"8ed4af2db63577128f6e1ea4b495e30dd14aa7379046fd821cf1c7e913e48d10"} Dec 10 19:02:18 crc kubenswrapper[4828]: I1210 19:02:18.839957 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.148907 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/thanos-querier-7bc849d657-tw8qx"] Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.151050 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.153353 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-metrics" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.153565 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-tls" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.153827 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.154009 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-dockercfg-gs9xv" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.154162 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-web" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.154270 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-rules" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.157044 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-grpc-tls-bhavm7fsdvavq" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.162285 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-7bc849d657-tw8qx"] Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.291790 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.291859 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7959a198-4584-426b-84ac-ffdc81bdeaf5-metrics-client-ca\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.291893 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.291914 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-thanos-querier-tls\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.291933 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-grpc-tls\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.291972 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.291992 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8dpw\" (UniqueName: \"kubernetes.io/projected/7959a198-4584-426b-84ac-ffdc81bdeaf5-kube-api-access-w8dpw\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.292036 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.393216 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.393267 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.393306 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7959a198-4584-426b-84ac-ffdc81bdeaf5-metrics-client-ca\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.393339 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.393360 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-thanos-querier-tls\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.393379 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-grpc-tls\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.393399 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.393418 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8dpw\" (UniqueName: \"kubernetes.io/projected/7959a198-4584-426b-84ac-ffdc81bdeaf5-kube-api-access-w8dpw\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.394772 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7959a198-4584-426b-84ac-ffdc81bdeaf5-metrics-client-ca\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.397774 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.398204 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.398367 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.398413 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-grpc-tls\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.398523 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-thanos-querier-tls\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.401179 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/7959a198-4584-426b-84ac-ffdc81bdeaf5-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.407530 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8dpw\" (UniqueName: \"kubernetes.io/projected/7959a198-4584-426b-84ac-ffdc81bdeaf5-kube-api-access-w8dpw\") pod \"thanos-querier-7bc849d657-tw8qx\" (UID: \"7959a198-4584-426b-84ac-ffdc81bdeaf5\") " pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.473095 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.661514 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"f8ab78da-de77-4dd1-9fd0-822314f0691e","Type":"ContainerStarted","Data":"8503e5cfe0215175206554dd9b7b152e557b24d85e8c212d744bab86a605b500"} Dec 10 19:02:19 crc kubenswrapper[4828]: I1210 19:02:19.758747 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-7bc849d657-tw8qx"] Dec 10 19:02:20 crc kubenswrapper[4828]: I1210 19:02:20.668626 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" event={"ID":"b067e07f-eb76-4de2-a145-e5a4d0770183","Type":"ContainerStarted","Data":"4e627447b8ed4ecffbb56ad09169706ae3b373933cb58cea9d89ad7d2d3ed5e9"} Dec 10 19:02:20 crc kubenswrapper[4828]: I1210 19:02:20.669001 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" event={"ID":"b067e07f-eb76-4de2-a145-e5a4d0770183","Type":"ContainerStarted","Data":"aae99e7f3390cace83bf8c2589025f68f1140abf4e7811b8aa9d9db2616335a3"} Dec 10 19:02:20 crc kubenswrapper[4828]: I1210 19:02:20.669014 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" event={"ID":"b067e07f-eb76-4de2-a145-e5a4d0770183","Type":"ContainerStarted","Data":"e947ffc12c5effa2c1abf5d3f254d17f30533eba18c37421804545d4367e8af9"} Dec 10 19:02:20 crc kubenswrapper[4828]: I1210 19:02:20.671714 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-nxjxg" event={"ID":"f47fe344-778e-4d41-aede-dd74a4c8d1a4","Type":"ContainerDied","Data":"9318b857ef253a456e7bef46eb7d3af824dfbde363c28df1a497e51b8dcae63a"} Dec 10 19:02:20 crc kubenswrapper[4828]: I1210 19:02:20.671611 4828 generic.go:334] "Generic (PLEG): container finished" podID="f47fe344-778e-4d41-aede-dd74a4c8d1a4" containerID="9318b857ef253a456e7bef46eb7d3af824dfbde363c28df1a497e51b8dcae63a" exitCode=0 Dec 10 19:02:20 crc kubenswrapper[4828]: I1210 19:02:20.674467 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" event={"ID":"7959a198-4584-426b-84ac-ffdc81bdeaf5","Type":"ContainerStarted","Data":"313db379b6f3a6004f9aac0164bd204a841a115f52373bf991e8a1b246b36a46"} Dec 10 19:02:20 crc kubenswrapper[4828]: I1210 19:02:20.688607 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-snw8w" podStartSLOduration=1.978512887 podStartE2EDuration="3.688586105s" podCreationTimestamp="2025-12-10 19:02:17 +0000 UTC" firstStartedPulling="2025-12-10 19:02:17.853031346 +0000 UTC m=+418.363642351" lastFinishedPulling="2025-12-10 19:02:19.563104564 +0000 UTC m=+420.073715569" observedRunningTime="2025-12-10 19:02:20.685264417 +0000 UTC m=+421.195875442" watchObservedRunningTime="2025-12-10 19:02:20.688586105 +0000 UTC m=+421.199197110" Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.230899 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.231241 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.231287 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.231931 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b8b8c9cc4d09570836d4fe54266d84e397b53b05d934a809d3077cafadbddea8"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.231992 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://b8b8c9cc4d09570836d4fe54266d84e397b53b05d934a809d3077cafadbddea8" gracePeriod=600 Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.696660 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="b8b8c9cc4d09570836d4fe54266d84e397b53b05d934a809d3077cafadbddea8" exitCode=0 Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.696725 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"b8b8c9cc4d09570836d4fe54266d84e397b53b05d934a809d3077cafadbddea8"} Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.696753 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"9e2e9020bcb1e1466a5a79b6ec9f630360e0d3e7e570a3abeb544c0a74e30c14"} Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.696769 4828 scope.go:117] "RemoveContainer" containerID="f85867e99620d4c30d8e9756a734963d76196ae2b0dafe3264aa126066672a71" Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.702450 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" event={"ID":"9841286a-46d2-4d2c-be70-834b9bcda856","Type":"ContainerStarted","Data":"bf6f9b11ceca6153b617dda12fb29eb148edff5e4b9151ad7f80f4c0845c88ce"} Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.708032 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-nxjxg" event={"ID":"f47fe344-778e-4d41-aede-dd74a4c8d1a4","Type":"ContainerStarted","Data":"90dea50b244432ad466ec98f9bfd7b67094ec1ea56c2a42e3303dde08c313e29"} Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.708096 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-nxjxg" event={"ID":"f47fe344-778e-4d41-aede-dd74a4c8d1a4","Type":"ContainerStarted","Data":"84bdad02009aa9fea5fa0e17db3d53f85deea6f7c8bed6e5d42a0a14ad9f976e"} Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.709851 4828 generic.go:334] "Generic (PLEG): container finished" podID="f8ab78da-de77-4dd1-9fd0-822314f0691e" containerID="e71b34cf771b6c6e6315159d38f411723f41335dfb44151703215daf9da92aa4" exitCode=0 Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.710433 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"f8ab78da-de77-4dd1-9fd0-822314f0691e","Type":"ContainerDied","Data":"e71b34cf771b6c6e6315159d38f411723f41335dfb44151703215daf9da92aa4"} Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.725603 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/openshift-state-metrics-566fddb674-wwqgv" podStartSLOduration=2.328672852 podStartE2EDuration="4.725584084s" podCreationTimestamp="2025-12-10 19:02:17 +0000 UTC" firstStartedPulling="2025-12-10 19:02:18.196077692 +0000 UTC m=+418.706688697" lastFinishedPulling="2025-12-10 19:02:20.592988904 +0000 UTC m=+421.103599929" observedRunningTime="2025-12-10 19:02:21.725054719 +0000 UTC m=+422.235665734" watchObservedRunningTime="2025-12-10 19:02:21.725584084 +0000 UTC m=+422.236195089" Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.747041 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-nxjxg" podStartSLOduration=3.227691076 podStartE2EDuration="4.747023275s" podCreationTimestamp="2025-12-10 19:02:17 +0000 UTC" firstStartedPulling="2025-12-10 19:02:18.045073793 +0000 UTC m=+418.555684808" lastFinishedPulling="2025-12-10 19:02:19.564406002 +0000 UTC m=+420.075017007" observedRunningTime="2025-12-10 19:02:21.746567271 +0000 UTC m=+422.257178276" watchObservedRunningTime="2025-12-10 19:02:21.747023275 +0000 UTC m=+422.257634280" Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.900401 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-74575c476-tbwqf"] Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.901113 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:21 crc kubenswrapper[4828]: I1210 19:02:21.949073 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-74575c476-tbwqf"] Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.030459 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfppc\" (UniqueName: \"kubernetes.io/projected/ab1e4960-2dca-4f0a-a717-d6efed72725e-kube-api-access-pfppc\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.030532 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-trusted-ca-bundle\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.030570 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-service-ca\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.030597 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-oauth-serving-cert\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.030630 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-oauth-config\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.030647 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-config\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.030723 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-serving-cert\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.131646 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-oauth-config\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.131702 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-config\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.131747 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-serving-cert\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.131817 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfppc\" (UniqueName: \"kubernetes.io/projected/ab1e4960-2dca-4f0a-a717-d6efed72725e-kube-api-access-pfppc\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.131839 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-trusted-ca-bundle\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.131870 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-service-ca\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.131895 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-oauth-serving-cert\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.132695 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-config\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.132782 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-oauth-serving-cert\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.132988 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-service-ca\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.133122 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-trusted-ca-bundle\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.138783 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-serving-cert\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.138846 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-oauth-config\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.154584 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfppc\" (UniqueName: \"kubernetes.io/projected/ab1e4960-2dca-4f0a-a717-d6efed72725e-kube-api-access-pfppc\") pod \"console-74575c476-tbwqf\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.221398 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.362465 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-64f9cfc58-dfdqs"] Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.363360 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.365473 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kubelet-serving-ca-bundle" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.365630 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-client-certs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.365965 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-tls" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.366187 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-server-audit-profiles" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.366295 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-8gj8f7sr8pclg" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.367097 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-dockercfg-plwss" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.374661 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-64f9cfc58-dfdqs"] Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.536401 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7d918d84-d473-471c-932a-2cc70e097cc9-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.536747 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/7d918d84-d473-471c-932a-2cc70e097cc9-metrics-server-audit-profiles\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.536776 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d918d84-d473-471c-932a-2cc70e097cc9-client-ca-bundle\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.536841 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq72w\" (UniqueName: \"kubernetes.io/projected/7d918d84-d473-471c-932a-2cc70e097cc9-kube-api-access-zq72w\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.536865 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/7d918d84-d473-471c-932a-2cc70e097cc9-secret-metrics-client-certs\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.536887 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/7d918d84-d473-471c-932a-2cc70e097cc9-secret-metrics-server-tls\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.536924 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/7d918d84-d473-471c-932a-2cc70e097cc9-audit-log\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.637752 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7d918d84-d473-471c-932a-2cc70e097cc9-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.637841 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/7d918d84-d473-471c-932a-2cc70e097cc9-metrics-server-audit-profiles\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.637871 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d918d84-d473-471c-932a-2cc70e097cc9-client-ca-bundle\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.637897 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq72w\" (UniqueName: \"kubernetes.io/projected/7d918d84-d473-471c-932a-2cc70e097cc9-kube-api-access-zq72w\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.637917 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/7d918d84-d473-471c-932a-2cc70e097cc9-secret-metrics-client-certs\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.637937 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/7d918d84-d473-471c-932a-2cc70e097cc9-secret-metrics-server-tls\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.637972 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/7d918d84-d473-471c-932a-2cc70e097cc9-audit-log\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.638563 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/7d918d84-d473-471c-932a-2cc70e097cc9-audit-log\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.638639 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7d918d84-d473-471c-932a-2cc70e097cc9-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.639920 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/7d918d84-d473-471c-932a-2cc70e097cc9-metrics-server-audit-profiles\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.645465 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/7d918d84-d473-471c-932a-2cc70e097cc9-secret-metrics-server-tls\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.645677 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/7d918d84-d473-471c-932a-2cc70e097cc9-secret-metrics-client-certs\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.646855 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d918d84-d473-471c-932a-2cc70e097cc9-client-ca-bundle\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.654158 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq72w\" (UniqueName: \"kubernetes.io/projected/7d918d84-d473-471c-932a-2cc70e097cc9-kube-api-access-zq72w\") pod \"metrics-server-64f9cfc58-dfdqs\" (UID: \"7d918d84-d473-471c-932a-2cc70e097cc9\") " pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.716744 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" event={"ID":"7959a198-4584-426b-84ac-ffdc81bdeaf5","Type":"ContainerStarted","Data":"8e04d7328660d9848feb0c2ca0adb6a9d0af208c8592b42b4d2fd365626273dc"} Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.716807 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" event={"ID":"7959a198-4584-426b-84ac-ffdc81bdeaf5","Type":"ContainerStarted","Data":"91ed798449c087bb9d59e27dd5aaeb76d31e14c2bbe092dc6221f88c98275fa0"} Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.729560 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.757116 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-74575c476-tbwqf"] Dec 10 19:02:22 crc kubenswrapper[4828]: W1210 19:02:22.761910 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab1e4960_2dca_4f0a_a717_d6efed72725e.slice/crio-71b7f4c097d4db5108111fd234edf55ce1ff5032126b8c73a8e85169558634b2 WatchSource:0}: Error finding container 71b7f4c097d4db5108111fd234edf55ce1ff5032126b8c73a8e85169558634b2: Status 404 returned error can't find the container with id 71b7f4c097d4db5108111fd234edf55ce1ff5032126b8c73a8e85169558634b2 Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.875696 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/monitoring-plugin-8448c556c8-ntsxp"] Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.876492 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-8448c556c8-ntsxp" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.878851 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"default-dockercfg-6tstp" Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.882883 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-8448c556c8-ntsxp"] Dec 10 19:02:22 crc kubenswrapper[4828]: I1210 19:02:22.885637 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"monitoring-plugin-cert" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.047851 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/13f52401-9278-44ab-a100-36ec7ad965e6-monitoring-plugin-cert\") pod \"monitoring-plugin-8448c556c8-ntsxp\" (UID: \"13f52401-9278-44ab-a100-36ec7ad965e6\") " pod="openshift-monitoring/monitoring-plugin-8448c556c8-ntsxp" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.149326 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/13f52401-9278-44ab-a100-36ec7ad965e6-monitoring-plugin-cert\") pod \"monitoring-plugin-8448c556c8-ntsxp\" (UID: \"13f52401-9278-44ab-a100-36ec7ad965e6\") " pod="openshift-monitoring/monitoring-plugin-8448c556c8-ntsxp" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.158441 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/13f52401-9278-44ab-a100-36ec7ad965e6-monitoring-plugin-cert\") pod \"monitoring-plugin-8448c556c8-ntsxp\" (UID: \"13f52401-9278-44ab-a100-36ec7ad965e6\") " pod="openshift-monitoring/monitoring-plugin-8448c556c8-ntsxp" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.164762 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-64f9cfc58-dfdqs"] Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.196106 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-8448c556c8-ntsxp" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.346469 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.348727 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.352336 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-sidecar-tls" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.352590 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-rbac-proxy" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.352831 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.353006 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-grpc-tls-dccpna08n3s25" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.353162 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-kube-rbac-proxy-web" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.353578 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-prometheus-http-client-file" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.354025 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls-assets-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.357766 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"serving-certs-ca-bundle" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.358030 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-dockercfg-tg72v" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.360455 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.363089 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-web-config" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.371936 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-k8s-rulefiles-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.381462 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-trusted-ca-bundle" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.389248 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.460620 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.460699 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.460736 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-config\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.460760 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.460879 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-config-out\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.460966 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.460992 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.461024 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.461071 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.461116 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr8np\" (UniqueName: \"kubernetes.io/projected/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-kube-api-access-jr8np\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.461148 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.461175 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.461266 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.461303 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.461344 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.461370 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.461402 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-web-config\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.461437 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.564887 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.564958 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.564985 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.565001 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.565022 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-web-config\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.565041 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.565061 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.565081 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.565101 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-config\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.565118 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.565141 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-config-out\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.565164 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.565179 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.565196 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.565214 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.565247 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr8np\" (UniqueName: \"kubernetes.io/projected/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-kube-api-access-jr8np\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.565264 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.565282 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.566059 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.566273 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.568725 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.570509 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-web-config\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.570758 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.570786 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.570954 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.571465 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.571541 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.571625 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.571781 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.574554 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.575430 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.575551 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.575620 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-config-out\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.576161 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.585834 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr8np\" (UniqueName: \"kubernetes.io/projected/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-kube-api-access-jr8np\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.593234 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c-config\") pod \"prometheus-k8s-0\" (UID: \"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.679717 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.728217 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" event={"ID":"7959a198-4584-426b-84ac-ffdc81bdeaf5","Type":"ContainerStarted","Data":"afbfe8ed165b2965664bebe85e4cca6451af2131fe1f0447c76b7e8a2573f67c"} Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.730268 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74575c476-tbwqf" event={"ID":"ab1e4960-2dca-4f0a-a717-d6efed72725e","Type":"ContainerStarted","Data":"a3b084a1cc7a13de53a1cce01dbc253f9642070eb20e115d1776ba116007a947"} Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.730301 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74575c476-tbwqf" event={"ID":"ab1e4960-2dca-4f0a-a717-d6efed72725e","Type":"ContainerStarted","Data":"71b7f4c097d4db5108111fd234edf55ce1ff5032126b8c73a8e85169558634b2"} Dec 10 19:02:23 crc kubenswrapper[4828]: I1210 19:02:23.749998 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-74575c476-tbwqf" podStartSLOduration=2.749983484 podStartE2EDuration="2.749983484s" podCreationTimestamp="2025-12-10 19:02:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:02:23.748580803 +0000 UTC m=+424.259191808" watchObservedRunningTime="2025-12-10 19:02:23.749983484 +0000 UTC m=+424.260594489" Dec 10 19:02:24 crc kubenswrapper[4828]: I1210 19:02:24.127260 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-8448c556c8-ntsxp"] Dec 10 19:02:24 crc kubenswrapper[4828]: W1210 19:02:24.133709 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13f52401_9278_44ab_a100_36ec7ad965e6.slice/crio-56f8b2917a1a48a72283c583831ea53c41bccb30ff8c96bf95ea50f35d83b0c6 WatchSource:0}: Error finding container 56f8b2917a1a48a72283c583831ea53c41bccb30ff8c96bf95ea50f35d83b0c6: Status 404 returned error can't find the container with id 56f8b2917a1a48a72283c583831ea53c41bccb30ff8c96bf95ea50f35d83b0c6 Dec 10 19:02:24 crc kubenswrapper[4828]: I1210 19:02:24.285747 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Dec 10 19:02:24 crc kubenswrapper[4828]: W1210 19:02:24.387817 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d32a5b0_6b3e_4d83_aba9_c3d2ac49127c.slice/crio-18435dea472cb089e843feada8f45c7d19d88b5e1b4be7c7c7ddd99fd60c6d20 WatchSource:0}: Error finding container 18435dea472cb089e843feada8f45c7d19d88b5e1b4be7c7c7ddd99fd60c6d20: Status 404 returned error can't find the container with id 18435dea472cb089e843feada8f45c7d19d88b5e1b4be7c7c7ddd99fd60c6d20 Dec 10 19:02:24 crc kubenswrapper[4828]: I1210 19:02:24.743883 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"f8ab78da-de77-4dd1-9fd0-822314f0691e","Type":"ContainerStarted","Data":"0ad3186a94dea1b0e056be61de873021b2e9b871292d259dcbfe78d53a2de084"} Dec 10 19:02:24 crc kubenswrapper[4828]: I1210 19:02:24.743935 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"f8ab78da-de77-4dd1-9fd0-822314f0691e","Type":"ContainerStarted","Data":"085ccc3cd77487cc2b755e12e9eaa9312f9a29ec43673f19d7b52a2e7d906f2c"} Dec 10 19:02:24 crc kubenswrapper[4828]: I1210 19:02:24.743957 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"f8ab78da-de77-4dd1-9fd0-822314f0691e","Type":"ContainerStarted","Data":"6afbd7e51d4186ac5b6b6fc2db4871915c2eb73c1673c76841f7e176f5342bd5"} Dec 10 19:02:24 crc kubenswrapper[4828]: I1210 19:02:24.745195 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" event={"ID":"7d918d84-d473-471c-932a-2cc70e097cc9","Type":"ContainerStarted","Data":"66c4534f365d11b6c7f143c77951f82f08c447bca1dd212f0d1a6e371330995d"} Dec 10 19:02:24 crc kubenswrapper[4828]: I1210 19:02:24.747316 4828 generic.go:334] "Generic (PLEG): container finished" podID="3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c" containerID="cfa3953ca381f571a04b2373a9af2e56c403e5f68fcb91b5279f0b13a22740e3" exitCode=0 Dec 10 19:02:24 crc kubenswrapper[4828]: I1210 19:02:24.747397 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c","Type":"ContainerDied","Data":"cfa3953ca381f571a04b2373a9af2e56c403e5f68fcb91b5279f0b13a22740e3"} Dec 10 19:02:24 crc kubenswrapper[4828]: I1210 19:02:24.747417 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c","Type":"ContainerStarted","Data":"18435dea472cb089e843feada8f45c7d19d88b5e1b4be7c7c7ddd99fd60c6d20"} Dec 10 19:02:24 crc kubenswrapper[4828]: I1210 19:02:24.748570 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-8448c556c8-ntsxp" event={"ID":"13f52401-9278-44ab-a100-36ec7ad965e6","Type":"ContainerStarted","Data":"56f8b2917a1a48a72283c583831ea53c41bccb30ff8c96bf95ea50f35d83b0c6"} Dec 10 19:02:24 crc kubenswrapper[4828]: I1210 19:02:24.754636 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" event={"ID":"7959a198-4584-426b-84ac-ffdc81bdeaf5","Type":"ContainerStarted","Data":"8808f34ea4e9978c788ffc09c2aedb9cc5d2c505b53b0c805143549825fcd0e5"} Dec 10 19:02:25 crc kubenswrapper[4828]: I1210 19:02:25.769550 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" event={"ID":"7959a198-4584-426b-84ac-ffdc81bdeaf5","Type":"ContainerStarted","Data":"7dfb3480c4fee07f95cb76a36ea15fc1108933d1ca79a2681f9e00dd683d1745"} Dec 10 19:02:25 crc kubenswrapper[4828]: I1210 19:02:25.770594 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:25 crc kubenswrapper[4828]: I1210 19:02:25.770616 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" event={"ID":"7959a198-4584-426b-84ac-ffdc81bdeaf5","Type":"ContainerStarted","Data":"62b9d89c1a5306424dcf9b1dbfb3d7b2c5c9e4f6eb091d472a0c0d4d0e2e3fd4"} Dec 10 19:02:25 crc kubenswrapper[4828]: I1210 19:02:25.774839 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"f8ab78da-de77-4dd1-9fd0-822314f0691e","Type":"ContainerStarted","Data":"0f03414b758b157773eb409e1621dd3e4bb008f6fc58870db84c987a9429a5ee"} Dec 10 19:02:25 crc kubenswrapper[4828]: I1210 19:02:25.774879 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"f8ab78da-de77-4dd1-9fd0-822314f0691e","Type":"ContainerStarted","Data":"848be6d79385b41c363041797258643d7f55cdb57fbfdffa0bace49fc91de920"} Dec 10 19:02:25 crc kubenswrapper[4828]: I1210 19:02:25.774890 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"f8ab78da-de77-4dd1-9fd0-822314f0691e","Type":"ContainerStarted","Data":"3013980e9a1488391c124981f47a2d9f2a46b8f3eb4a289d133158f135dd7374"} Dec 10 19:02:25 crc kubenswrapper[4828]: I1210 19:02:25.790549 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" podStartSLOduration=2.120140653 podStartE2EDuration="6.790534589s" podCreationTimestamp="2025-12-10 19:02:19 +0000 UTC" firstStartedPulling="2025-12-10 19:02:19.763195347 +0000 UTC m=+420.273806352" lastFinishedPulling="2025-12-10 19:02:24.433589283 +0000 UTC m=+424.944200288" observedRunningTime="2025-12-10 19:02:25.78955863 +0000 UTC m=+426.300169635" watchObservedRunningTime="2025-12-10 19:02:25.790534589 +0000 UTC m=+426.301145584" Dec 10 19:02:25 crc kubenswrapper[4828]: I1210 19:02:25.821176 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=2.77171782 podStartE2EDuration="7.82115958s" podCreationTimestamp="2025-12-10 19:02:18 +0000 UTC" firstStartedPulling="2025-12-10 19:02:18.853401738 +0000 UTC m=+419.364012743" lastFinishedPulling="2025-12-10 19:02:23.902843498 +0000 UTC m=+424.413454503" observedRunningTime="2025-12-10 19:02:25.818179312 +0000 UTC m=+426.328790327" watchObservedRunningTime="2025-12-10 19:02:25.82115958 +0000 UTC m=+426.331770575" Dec 10 19:02:26 crc kubenswrapper[4828]: I1210 19:02:26.781881 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-8448c556c8-ntsxp" event={"ID":"13f52401-9278-44ab-a100-36ec7ad965e6","Type":"ContainerStarted","Data":"52f28f433ab9f70640031557edba45322e03332b9074e4a458d0486ea2bd718b"} Dec 10 19:02:26 crc kubenswrapper[4828]: I1210 19:02:26.783747 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/monitoring-plugin-8448c556c8-ntsxp" Dec 10 19:02:26 crc kubenswrapper[4828]: I1210 19:02:26.787778 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" event={"ID":"7d918d84-d473-471c-932a-2cc70e097cc9","Type":"ContainerStarted","Data":"d260468e85ba54875167594209c7c7902174f2b233d5191e873edd0fb5a900c9"} Dec 10 19:02:26 crc kubenswrapper[4828]: I1210 19:02:26.789671 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-8448c556c8-ntsxp" Dec 10 19:02:26 crc kubenswrapper[4828]: I1210 19:02:26.809343 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/monitoring-plugin-8448c556c8-ntsxp" podStartSLOduration=2.9817786 podStartE2EDuration="4.809326982s" podCreationTimestamp="2025-12-10 19:02:22 +0000 UTC" firstStartedPulling="2025-12-10 19:02:24.13590237 +0000 UTC m=+424.646513375" lastFinishedPulling="2025-12-10 19:02:25.963450752 +0000 UTC m=+426.474061757" observedRunningTime="2025-12-10 19:02:26.794553548 +0000 UTC m=+427.305164563" watchObservedRunningTime="2025-12-10 19:02:26.809326982 +0000 UTC m=+427.319937987" Dec 10 19:02:26 crc kubenswrapper[4828]: I1210 19:02:26.826594 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" podStartSLOduration=2.698382558 podStartE2EDuration="4.826573769s" podCreationTimestamp="2025-12-10 19:02:22 +0000 UTC" firstStartedPulling="2025-12-10 19:02:23.84099991 +0000 UTC m=+424.351610915" lastFinishedPulling="2025-12-10 19:02:25.969191121 +0000 UTC m=+426.479802126" observedRunningTime="2025-12-10 19:02:26.818864513 +0000 UTC m=+427.329475518" watchObservedRunningTime="2025-12-10 19:02:26.826573769 +0000 UTC m=+427.337184774" Dec 10 19:02:28 crc kubenswrapper[4828]: I1210 19:02:28.804583 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c","Type":"ContainerStarted","Data":"3d886e6e96c2c18b141dc7eb6776d926d86b080a5b1c575959c0752037dbcbaa"} Dec 10 19:02:28 crc kubenswrapper[4828]: I1210 19:02:28.804969 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c","Type":"ContainerStarted","Data":"871dcf18431c2f03b2e8f0f1a8c0eab336b8e34c3cd75a07fa233afd95244c53"} Dec 10 19:02:29 crc kubenswrapper[4828]: I1210 19:02:29.483949 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-7bc849d657-tw8qx" Dec 10 19:02:29 crc kubenswrapper[4828]: I1210 19:02:29.812447 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c","Type":"ContainerStarted","Data":"1857d319569cdcb6626a9ff03c0f02e9ea376b2d2c569ae72d511f770f5f85a4"} Dec 10 19:02:29 crc kubenswrapper[4828]: I1210 19:02:29.812733 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c","Type":"ContainerStarted","Data":"2456d1bc6d91ace24de45da071c4837a45a4cdc6862626ef135d0ce813d95f4d"} Dec 10 19:02:29 crc kubenswrapper[4828]: I1210 19:02:29.812743 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c","Type":"ContainerStarted","Data":"32383a8e6de3b85069ca4cd6bfb4e5682fbd40c4ac63845d6a3be7b629569433"} Dec 10 19:02:29 crc kubenswrapper[4828]: I1210 19:02:29.812752 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"3d32a5b0-6b3e-4d83-aba9-c3d2ac49127c","Type":"ContainerStarted","Data":"fc739b918ba89ed22a2e545444aa2d993f7b65aafa336a5938ed9ae23878fb2b"} Dec 10 19:02:29 crc kubenswrapper[4828]: I1210 19:02:29.838986 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=3.063311397 podStartE2EDuration="6.838971546s" podCreationTimestamp="2025-12-10 19:02:23 +0000 UTC" firstStartedPulling="2025-12-10 19:02:24.749880342 +0000 UTC m=+425.260491347" lastFinishedPulling="2025-12-10 19:02:28.525540491 +0000 UTC m=+429.036151496" observedRunningTime="2025-12-10 19:02:29.836537015 +0000 UTC m=+430.347148030" watchObservedRunningTime="2025-12-10 19:02:29.838971546 +0000 UTC m=+430.349582551" Dec 10 19:02:32 crc kubenswrapper[4828]: I1210 19:02:32.221760 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:32 crc kubenswrapper[4828]: I1210 19:02:32.222209 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:32 crc kubenswrapper[4828]: I1210 19:02:32.226174 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:32 crc kubenswrapper[4828]: I1210 19:02:32.849895 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:02:32 crc kubenswrapper[4828]: I1210 19:02:32.902888 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-m4fkj"] Dec 10 19:02:33 crc kubenswrapper[4828]: I1210 19:02:33.680385 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:02:42 crc kubenswrapper[4828]: I1210 19:02:42.730241 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:42 crc kubenswrapper[4828]: I1210 19:02:42.731023 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:02:57 crc kubenswrapper[4828]: I1210 19:02:57.944114 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-m4fkj" podUID="a2629f52-3b35-4707-8a24-53d7bbb70465" containerName="console" containerID="cri-o://df082dd35197b2e6981657555c2e16a44917cd58def58e905691e03394e43d5c" gracePeriod=15 Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.300506 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-m4fkj_a2629f52-3b35-4707-8a24-53d7bbb70465/console/0.log" Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.300952 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.498551 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-trusted-ca-bundle\") pod \"a2629f52-3b35-4707-8a24-53d7bbb70465\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.498609 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-service-ca\") pod \"a2629f52-3b35-4707-8a24-53d7bbb70465\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.498678 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a2629f52-3b35-4707-8a24-53d7bbb70465-console-oauth-config\") pod \"a2629f52-3b35-4707-8a24-53d7bbb70465\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.498746 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgvzq\" (UniqueName: \"kubernetes.io/projected/a2629f52-3b35-4707-8a24-53d7bbb70465-kube-api-access-vgvzq\") pod \"a2629f52-3b35-4707-8a24-53d7bbb70465\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.498819 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-console-config\") pod \"a2629f52-3b35-4707-8a24-53d7bbb70465\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.498864 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a2629f52-3b35-4707-8a24-53d7bbb70465-console-serving-cert\") pod \"a2629f52-3b35-4707-8a24-53d7bbb70465\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.498912 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-oauth-serving-cert\") pod \"a2629f52-3b35-4707-8a24-53d7bbb70465\" (UID: \"a2629f52-3b35-4707-8a24-53d7bbb70465\") " Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.499587 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-service-ca" (OuterVolumeSpecName: "service-ca") pod "a2629f52-3b35-4707-8a24-53d7bbb70465" (UID: "a2629f52-3b35-4707-8a24-53d7bbb70465"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.499618 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "a2629f52-3b35-4707-8a24-53d7bbb70465" (UID: "a2629f52-3b35-4707-8a24-53d7bbb70465"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.499615 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "a2629f52-3b35-4707-8a24-53d7bbb70465" (UID: "a2629f52-3b35-4707-8a24-53d7bbb70465"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.499695 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-console-config" (OuterVolumeSpecName: "console-config") pod "a2629f52-3b35-4707-8a24-53d7bbb70465" (UID: "a2629f52-3b35-4707-8a24-53d7bbb70465"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.504173 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2629f52-3b35-4707-8a24-53d7bbb70465-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "a2629f52-3b35-4707-8a24-53d7bbb70465" (UID: "a2629f52-3b35-4707-8a24-53d7bbb70465"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.504551 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2629f52-3b35-4707-8a24-53d7bbb70465-kube-api-access-vgvzq" (OuterVolumeSpecName: "kube-api-access-vgvzq") pod "a2629f52-3b35-4707-8a24-53d7bbb70465" (UID: "a2629f52-3b35-4707-8a24-53d7bbb70465"). InnerVolumeSpecName "kube-api-access-vgvzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.507779 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2629f52-3b35-4707-8a24-53d7bbb70465-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "a2629f52-3b35-4707-8a24-53d7bbb70465" (UID: "a2629f52-3b35-4707-8a24-53d7bbb70465"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.600231 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgvzq\" (UniqueName: \"kubernetes.io/projected/a2629f52-3b35-4707-8a24-53d7bbb70465-kube-api-access-vgvzq\") on node \"crc\" DevicePath \"\"" Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.600532 4828 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-console-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.600542 4828 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a2629f52-3b35-4707-8a24-53d7bbb70465-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.600551 4828 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.600560 4828 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.600569 4828 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a2629f52-3b35-4707-8a24-53d7bbb70465-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 19:02:58 crc kubenswrapper[4828]: I1210 19:02:58.600578 4828 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a2629f52-3b35-4707-8a24-53d7bbb70465-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:02:59 crc kubenswrapper[4828]: I1210 19:02:59.018750 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-m4fkj_a2629f52-3b35-4707-8a24-53d7bbb70465/console/0.log" Dec 10 19:02:59 crc kubenswrapper[4828]: I1210 19:02:59.018897 4828 generic.go:334] "Generic (PLEG): container finished" podID="a2629f52-3b35-4707-8a24-53d7bbb70465" containerID="df082dd35197b2e6981657555c2e16a44917cd58def58e905691e03394e43d5c" exitCode=2 Dec 10 19:02:59 crc kubenswrapper[4828]: I1210 19:02:59.018961 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m4fkj" event={"ID":"a2629f52-3b35-4707-8a24-53d7bbb70465","Type":"ContainerDied","Data":"df082dd35197b2e6981657555c2e16a44917cd58def58e905691e03394e43d5c"} Dec 10 19:02:59 crc kubenswrapper[4828]: I1210 19:02:59.019064 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m4fkj" event={"ID":"a2629f52-3b35-4707-8a24-53d7bbb70465","Type":"ContainerDied","Data":"bd51afe045f8dfaf1d9d491409d261c81fe127e419b1041386caf3c0511323f1"} Dec 10 19:02:59 crc kubenswrapper[4828]: I1210 19:02:59.019093 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m4fkj" Dec 10 19:02:59 crc kubenswrapper[4828]: I1210 19:02:59.019106 4828 scope.go:117] "RemoveContainer" containerID="df082dd35197b2e6981657555c2e16a44917cd58def58e905691e03394e43d5c" Dec 10 19:02:59 crc kubenswrapper[4828]: I1210 19:02:59.053593 4828 scope.go:117] "RemoveContainer" containerID="df082dd35197b2e6981657555c2e16a44917cd58def58e905691e03394e43d5c" Dec 10 19:02:59 crc kubenswrapper[4828]: E1210 19:02:59.054247 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df082dd35197b2e6981657555c2e16a44917cd58def58e905691e03394e43d5c\": container with ID starting with df082dd35197b2e6981657555c2e16a44917cd58def58e905691e03394e43d5c not found: ID does not exist" containerID="df082dd35197b2e6981657555c2e16a44917cd58def58e905691e03394e43d5c" Dec 10 19:02:59 crc kubenswrapper[4828]: I1210 19:02:59.054374 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df082dd35197b2e6981657555c2e16a44917cd58def58e905691e03394e43d5c"} err="failed to get container status \"df082dd35197b2e6981657555c2e16a44917cd58def58e905691e03394e43d5c\": rpc error: code = NotFound desc = could not find container \"df082dd35197b2e6981657555c2e16a44917cd58def58e905691e03394e43d5c\": container with ID starting with df082dd35197b2e6981657555c2e16a44917cd58def58e905691e03394e43d5c not found: ID does not exist" Dec 10 19:02:59 crc kubenswrapper[4828]: I1210 19:02:59.063684 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-m4fkj"] Dec 10 19:02:59 crc kubenswrapper[4828]: I1210 19:02:59.074168 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-m4fkj"] Dec 10 19:02:59 crc kubenswrapper[4828]: I1210 19:02:59.802890 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2629f52-3b35-4707-8a24-53d7bbb70465" path="/var/lib/kubelet/pods/a2629f52-3b35-4707-8a24-53d7bbb70465/volumes" Dec 10 19:03:02 crc kubenswrapper[4828]: I1210 19:03:02.735605 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:03:02 crc kubenswrapper[4828]: I1210 19:03:02.740890 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-64f9cfc58-dfdqs" Dec 10 19:03:23 crc kubenswrapper[4828]: I1210 19:03:23.680438 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:03:23 crc kubenswrapper[4828]: I1210 19:03:23.723116 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:03:24 crc kubenswrapper[4828]: I1210 19:03:24.205475 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.791930 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-8486f475fd-6jrxr"] Dec 10 19:03:50 crc kubenswrapper[4828]: E1210 19:03:50.794007 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2629f52-3b35-4707-8a24-53d7bbb70465" containerName="console" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.794107 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2629f52-3b35-4707-8a24-53d7bbb70465" containerName="console" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.794323 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2629f52-3b35-4707-8a24-53d7bbb70465" containerName="console" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.794966 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.814315 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-8486f475fd-6jrxr"] Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.849265 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-service-ca\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.849622 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-oauth-config\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.849731 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-trusted-ca-bundle\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.849839 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-config\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.849948 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-oauth-serving-cert\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.850101 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpmfk\" (UniqueName: \"kubernetes.io/projected/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-kube-api-access-hpmfk\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.850634 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-serving-cert\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.952637 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-config\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.952921 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-oauth-serving-cert\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.953042 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpmfk\" (UniqueName: \"kubernetes.io/projected/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-kube-api-access-hpmfk\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.953139 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-serving-cert\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.953228 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-service-ca\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.953366 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-oauth-config\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.953460 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-trusted-ca-bundle\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.953648 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-config\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.953738 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-oauth-serving-cert\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.954793 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-service-ca\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.955012 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-trusted-ca-bundle\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.958203 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-oauth-config\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.958470 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-serving-cert\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:50 crc kubenswrapper[4828]: I1210 19:03:50.970640 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpmfk\" (UniqueName: \"kubernetes.io/projected/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-kube-api-access-hpmfk\") pod \"console-8486f475fd-6jrxr\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:51 crc kubenswrapper[4828]: I1210 19:03:51.124133 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:03:51 crc kubenswrapper[4828]: I1210 19:03:51.502469 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-8486f475fd-6jrxr"] Dec 10 19:03:52 crc kubenswrapper[4828]: I1210 19:03:52.377285 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8486f475fd-6jrxr" event={"ID":"434cf9e4-da1b-49cb-a641-9a84ebe2e22c","Type":"ContainerStarted","Data":"2e35b4ad0b8cb77ea24d1d1476570efc4ac30280c569dc13dfaf10b670724d9b"} Dec 10 19:03:52 crc kubenswrapper[4828]: I1210 19:03:52.377597 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8486f475fd-6jrxr" event={"ID":"434cf9e4-da1b-49cb-a641-9a84ebe2e22c","Type":"ContainerStarted","Data":"04f8d8d109b5226c8e13e3b971167463ebca67ee078e1c428abe21f67f85b568"} Dec 10 19:03:52 crc kubenswrapper[4828]: I1210 19:03:52.399755 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-8486f475fd-6jrxr" podStartSLOduration=2.399737372 podStartE2EDuration="2.399737372s" podCreationTimestamp="2025-12-10 19:03:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:03:52.395601191 +0000 UTC m=+512.906212226" watchObservedRunningTime="2025-12-10 19:03:52.399737372 +0000 UTC m=+512.910348367" Dec 10 19:04:01 crc kubenswrapper[4828]: I1210 19:04:01.125196 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:04:01 crc kubenswrapper[4828]: I1210 19:04:01.125546 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:04:01 crc kubenswrapper[4828]: I1210 19:04:01.129279 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:04:01 crc kubenswrapper[4828]: I1210 19:04:01.427062 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:04:01 crc kubenswrapper[4828]: I1210 19:04:01.481527 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-74575c476-tbwqf"] Dec 10 19:04:21 crc kubenswrapper[4828]: I1210 19:04:21.230722 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:04:21 crc kubenswrapper[4828]: I1210 19:04:21.231949 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:04:23 crc kubenswrapper[4828]: I1210 19:04:23.972327 4828 scope.go:117] "RemoveContainer" containerID="be827b35061551e27e2ea93f8ae82720cf53c1222786b1adcdb55d1302abee78" Dec 10 19:04:26 crc kubenswrapper[4828]: I1210 19:04:26.523147 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-74575c476-tbwqf" podUID="ab1e4960-2dca-4f0a-a717-d6efed72725e" containerName="console" containerID="cri-o://a3b084a1cc7a13de53a1cce01dbc253f9642070eb20e115d1776ba116007a947" gracePeriod=15 Dec 10 19:04:26 crc kubenswrapper[4828]: I1210 19:04:26.950364 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-74575c476-tbwqf_ab1e4960-2dca-4f0a-a717-d6efed72725e/console/0.log" Dec 10 19:04:26 crc kubenswrapper[4828]: I1210 19:04:26.950672 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.046194 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-oauth-serving-cert\") pod \"ab1e4960-2dca-4f0a-a717-d6efed72725e\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.046234 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-config\") pod \"ab1e4960-2dca-4f0a-a717-d6efed72725e\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.046268 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-oauth-config\") pod \"ab1e4960-2dca-4f0a-a717-d6efed72725e\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.046317 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-service-ca\") pod \"ab1e4960-2dca-4f0a-a717-d6efed72725e\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.046335 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-trusted-ca-bundle\") pod \"ab1e4960-2dca-4f0a-a717-d6efed72725e\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.046354 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfppc\" (UniqueName: \"kubernetes.io/projected/ab1e4960-2dca-4f0a-a717-d6efed72725e-kube-api-access-pfppc\") pod \"ab1e4960-2dca-4f0a-a717-d6efed72725e\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.046371 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-serving-cert\") pod \"ab1e4960-2dca-4f0a-a717-d6efed72725e\" (UID: \"ab1e4960-2dca-4f0a-a717-d6efed72725e\") " Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.047480 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "ab1e4960-2dca-4f0a-a717-d6efed72725e" (UID: "ab1e4960-2dca-4f0a-a717-d6efed72725e"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.047681 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-config" (OuterVolumeSpecName: "console-config") pod "ab1e4960-2dca-4f0a-a717-d6efed72725e" (UID: "ab1e4960-2dca-4f0a-a717-d6efed72725e"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.047920 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-service-ca" (OuterVolumeSpecName: "service-ca") pod "ab1e4960-2dca-4f0a-a717-d6efed72725e" (UID: "ab1e4960-2dca-4f0a-a717-d6efed72725e"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.048438 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "ab1e4960-2dca-4f0a-a717-d6efed72725e" (UID: "ab1e4960-2dca-4f0a-a717-d6efed72725e"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.050938 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "ab1e4960-2dca-4f0a-a717-d6efed72725e" (UID: "ab1e4960-2dca-4f0a-a717-d6efed72725e"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.051897 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab1e4960-2dca-4f0a-a717-d6efed72725e-kube-api-access-pfppc" (OuterVolumeSpecName: "kube-api-access-pfppc") pod "ab1e4960-2dca-4f0a-a717-d6efed72725e" (UID: "ab1e4960-2dca-4f0a-a717-d6efed72725e"). InnerVolumeSpecName "kube-api-access-pfppc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.052042 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "ab1e4960-2dca-4f0a-a717-d6efed72725e" (UID: "ab1e4960-2dca-4f0a-a717-d6efed72725e"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.147851 4828 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.147882 4828 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.147893 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfppc\" (UniqueName: \"kubernetes.io/projected/ab1e4960-2dca-4f0a-a717-d6efed72725e-kube-api-access-pfppc\") on node \"crc\" DevicePath \"\"" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.147905 4828 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.147915 4828 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.147922 4828 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.147929 4828 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ab1e4960-2dca-4f0a-a717-d6efed72725e-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.598095 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-74575c476-tbwqf_ab1e4960-2dca-4f0a-a717-d6efed72725e/console/0.log" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.598196 4828 generic.go:334] "Generic (PLEG): container finished" podID="ab1e4960-2dca-4f0a-a717-d6efed72725e" containerID="a3b084a1cc7a13de53a1cce01dbc253f9642070eb20e115d1776ba116007a947" exitCode=2 Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.598257 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74575c476-tbwqf" event={"ID":"ab1e4960-2dca-4f0a-a717-d6efed72725e","Type":"ContainerDied","Data":"a3b084a1cc7a13de53a1cce01dbc253f9642070eb20e115d1776ba116007a947"} Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.598306 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74575c476-tbwqf" event={"ID":"ab1e4960-2dca-4f0a-a717-d6efed72725e","Type":"ContainerDied","Data":"71b7f4c097d4db5108111fd234edf55ce1ff5032126b8c73a8e85169558634b2"} Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.598347 4828 scope.go:117] "RemoveContainer" containerID="a3b084a1cc7a13de53a1cce01dbc253f9642070eb20e115d1776ba116007a947" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.598453 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74575c476-tbwqf" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.624579 4828 scope.go:117] "RemoveContainer" containerID="a3b084a1cc7a13de53a1cce01dbc253f9642070eb20e115d1776ba116007a947" Dec 10 19:04:27 crc kubenswrapper[4828]: E1210 19:04:27.625259 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3b084a1cc7a13de53a1cce01dbc253f9642070eb20e115d1776ba116007a947\": container with ID starting with a3b084a1cc7a13de53a1cce01dbc253f9642070eb20e115d1776ba116007a947 not found: ID does not exist" containerID="a3b084a1cc7a13de53a1cce01dbc253f9642070eb20e115d1776ba116007a947" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.625318 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3b084a1cc7a13de53a1cce01dbc253f9642070eb20e115d1776ba116007a947"} err="failed to get container status \"a3b084a1cc7a13de53a1cce01dbc253f9642070eb20e115d1776ba116007a947\": rpc error: code = NotFound desc = could not find container \"a3b084a1cc7a13de53a1cce01dbc253f9642070eb20e115d1776ba116007a947\": container with ID starting with a3b084a1cc7a13de53a1cce01dbc253f9642070eb20e115d1776ba116007a947 not found: ID does not exist" Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.638888 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-74575c476-tbwqf"] Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.643915 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-74575c476-tbwqf"] Dec 10 19:04:27 crc kubenswrapper[4828]: I1210 19:04:27.800364 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab1e4960-2dca-4f0a-a717-d6efed72725e" path="/var/lib/kubelet/pods/ab1e4960-2dca-4f0a-a717-d6efed72725e/volumes" Dec 10 19:04:51 crc kubenswrapper[4828]: I1210 19:04:51.231132 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:04:51 crc kubenswrapper[4828]: I1210 19:04:51.232049 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:05:21 crc kubenswrapper[4828]: I1210 19:05:21.230352 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:05:21 crc kubenswrapper[4828]: I1210 19:05:21.231199 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:05:21 crc kubenswrapper[4828]: I1210 19:05:21.231269 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 19:05:21 crc kubenswrapper[4828]: I1210 19:05:21.232256 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9e2e9020bcb1e1466a5a79b6ec9f630360e0d3e7e570a3abeb544c0a74e30c14"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:05:21 crc kubenswrapper[4828]: I1210 19:05:21.232369 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://9e2e9020bcb1e1466a5a79b6ec9f630360e0d3e7e570a3abeb544c0a74e30c14" gracePeriod=600 Dec 10 19:05:21 crc kubenswrapper[4828]: I1210 19:05:21.986645 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="9e2e9020bcb1e1466a5a79b6ec9f630360e0d3e7e570a3abeb544c0a74e30c14" exitCode=0 Dec 10 19:05:21 crc kubenswrapper[4828]: I1210 19:05:21.986723 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"9e2e9020bcb1e1466a5a79b6ec9f630360e0d3e7e570a3abeb544c0a74e30c14"} Dec 10 19:05:21 crc kubenswrapper[4828]: I1210 19:05:21.987313 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"1d400e5fe2702c739b92ea52b2e9453f005b08c30f146bdcec9d4d95ba2cb93b"} Dec 10 19:05:21 crc kubenswrapper[4828]: I1210 19:05:21.987341 4828 scope.go:117] "RemoveContainer" containerID="b8b8c9cc4d09570836d4fe54266d84e397b53b05d934a809d3077cafadbddea8" Dec 10 19:05:24 crc kubenswrapper[4828]: I1210 19:05:24.014580 4828 scope.go:117] "RemoveContainer" containerID="1c65abebabfa38770059d8a3ee3bad6c5cafea58415c02a7eaf3efcbbdbb797c" Dec 10 19:07:10 crc kubenswrapper[4828]: I1210 19:07:10.759860 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f"] Dec 10 19:07:10 crc kubenswrapper[4828]: E1210 19:07:10.760724 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab1e4960-2dca-4f0a-a717-d6efed72725e" containerName="console" Dec 10 19:07:10 crc kubenswrapper[4828]: I1210 19:07:10.760747 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab1e4960-2dca-4f0a-a717-d6efed72725e" containerName="console" Dec 10 19:07:10 crc kubenswrapper[4828]: I1210 19:07:10.760964 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab1e4960-2dca-4f0a-a717-d6efed72725e" containerName="console" Dec 10 19:07:10 crc kubenswrapper[4828]: I1210 19:07:10.762167 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" Dec 10 19:07:10 crc kubenswrapper[4828]: I1210 19:07:10.764023 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 10 19:07:10 crc kubenswrapper[4828]: I1210 19:07:10.769175 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f"] Dec 10 19:07:10 crc kubenswrapper[4828]: I1210 19:07:10.869517 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhddw\" (UniqueName: \"kubernetes.io/projected/025427dc-d729-4b53-b665-de357c409d97-kube-api-access-mhddw\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f\" (UID: \"025427dc-d729-4b53-b665-de357c409d97\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" Dec 10 19:07:10 crc kubenswrapper[4828]: I1210 19:07:10.869626 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/025427dc-d729-4b53-b665-de357c409d97-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f\" (UID: \"025427dc-d729-4b53-b665-de357c409d97\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" Dec 10 19:07:10 crc kubenswrapper[4828]: I1210 19:07:10.869677 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/025427dc-d729-4b53-b665-de357c409d97-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f\" (UID: \"025427dc-d729-4b53-b665-de357c409d97\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" Dec 10 19:07:10 crc kubenswrapper[4828]: I1210 19:07:10.970819 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/025427dc-d729-4b53-b665-de357c409d97-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f\" (UID: \"025427dc-d729-4b53-b665-de357c409d97\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" Dec 10 19:07:10 crc kubenswrapper[4828]: I1210 19:07:10.971263 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/025427dc-d729-4b53-b665-de357c409d97-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f\" (UID: \"025427dc-d729-4b53-b665-de357c409d97\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" Dec 10 19:07:10 crc kubenswrapper[4828]: I1210 19:07:10.971467 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhddw\" (UniqueName: \"kubernetes.io/projected/025427dc-d729-4b53-b665-de357c409d97-kube-api-access-mhddw\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f\" (UID: \"025427dc-d729-4b53-b665-de357c409d97\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" Dec 10 19:07:10 crc kubenswrapper[4828]: I1210 19:07:10.971565 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/025427dc-d729-4b53-b665-de357c409d97-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f\" (UID: \"025427dc-d729-4b53-b665-de357c409d97\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" Dec 10 19:07:10 crc kubenswrapper[4828]: I1210 19:07:10.971830 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/025427dc-d729-4b53-b665-de357c409d97-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f\" (UID: \"025427dc-d729-4b53-b665-de357c409d97\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" Dec 10 19:07:10 crc kubenswrapper[4828]: I1210 19:07:10.993341 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhddw\" (UniqueName: \"kubernetes.io/projected/025427dc-d729-4b53-b665-de357c409d97-kube-api-access-mhddw\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f\" (UID: \"025427dc-d729-4b53-b665-de357c409d97\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" Dec 10 19:07:11 crc kubenswrapper[4828]: I1210 19:07:11.079309 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" Dec 10 19:07:11 crc kubenswrapper[4828]: I1210 19:07:11.265137 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f"] Dec 10 19:07:11 crc kubenswrapper[4828]: I1210 19:07:11.694982 4828 generic.go:334] "Generic (PLEG): container finished" podID="025427dc-d729-4b53-b665-de357c409d97" containerID="6b5a0df994dabfca7d9f986e0fe550e1ef3aa28db8d348b7559c291559002f21" exitCode=0 Dec 10 19:07:11 crc kubenswrapper[4828]: I1210 19:07:11.695068 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" event={"ID":"025427dc-d729-4b53-b665-de357c409d97","Type":"ContainerDied","Data":"6b5a0df994dabfca7d9f986e0fe550e1ef3aa28db8d348b7559c291559002f21"} Dec 10 19:07:11 crc kubenswrapper[4828]: I1210 19:07:11.695358 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" event={"ID":"025427dc-d729-4b53-b665-de357c409d97","Type":"ContainerStarted","Data":"56e7241d385a7cd8acd4b0cee4819b4fbf4223ee72b01142f709bb05077ddba1"} Dec 10 19:07:11 crc kubenswrapper[4828]: I1210 19:07:11.697586 4828 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:07:16 crc kubenswrapper[4828]: I1210 19:07:16.733203 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" event={"ID":"025427dc-d729-4b53-b665-de357c409d97","Type":"ContainerStarted","Data":"2f6266ee68b62d0acdd482fe1a03362c23c01149f73de3b1d42847b200a861e7"} Dec 10 19:07:17 crc kubenswrapper[4828]: I1210 19:07:17.743992 4828 generic.go:334] "Generic (PLEG): container finished" podID="025427dc-d729-4b53-b665-de357c409d97" containerID="2f6266ee68b62d0acdd482fe1a03362c23c01149f73de3b1d42847b200a861e7" exitCode=0 Dec 10 19:07:17 crc kubenswrapper[4828]: I1210 19:07:17.744087 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" event={"ID":"025427dc-d729-4b53-b665-de357c409d97","Type":"ContainerDied","Data":"2f6266ee68b62d0acdd482fe1a03362c23c01149f73de3b1d42847b200a861e7"} Dec 10 19:07:18 crc kubenswrapper[4828]: I1210 19:07:18.752490 4828 generic.go:334] "Generic (PLEG): container finished" podID="025427dc-d729-4b53-b665-de357c409d97" containerID="f12df01adec8ed7eb218049f0a6c914f1f702404f9026ad983736e277b9ae565" exitCode=0 Dec 10 19:07:18 crc kubenswrapper[4828]: I1210 19:07:18.752668 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" event={"ID":"025427dc-d729-4b53-b665-de357c409d97","Type":"ContainerDied","Data":"f12df01adec8ed7eb218049f0a6c914f1f702404f9026ad983736e277b9ae565"} Dec 10 19:07:20 crc kubenswrapper[4828]: I1210 19:07:20.062062 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" Dec 10 19:07:20 crc kubenswrapper[4828]: I1210 19:07:20.206085 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/025427dc-d729-4b53-b665-de357c409d97-bundle\") pod \"025427dc-d729-4b53-b665-de357c409d97\" (UID: \"025427dc-d729-4b53-b665-de357c409d97\") " Dec 10 19:07:20 crc kubenswrapper[4828]: I1210 19:07:20.206251 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/025427dc-d729-4b53-b665-de357c409d97-util\") pod \"025427dc-d729-4b53-b665-de357c409d97\" (UID: \"025427dc-d729-4b53-b665-de357c409d97\") " Dec 10 19:07:20 crc kubenswrapper[4828]: I1210 19:07:20.206320 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhddw\" (UniqueName: \"kubernetes.io/projected/025427dc-d729-4b53-b665-de357c409d97-kube-api-access-mhddw\") pod \"025427dc-d729-4b53-b665-de357c409d97\" (UID: \"025427dc-d729-4b53-b665-de357c409d97\") " Dec 10 19:07:20 crc kubenswrapper[4828]: I1210 19:07:20.210492 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/025427dc-d729-4b53-b665-de357c409d97-bundle" (OuterVolumeSpecName: "bundle") pod "025427dc-d729-4b53-b665-de357c409d97" (UID: "025427dc-d729-4b53-b665-de357c409d97"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:07:20 crc kubenswrapper[4828]: I1210 19:07:20.212437 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/025427dc-d729-4b53-b665-de357c409d97-kube-api-access-mhddw" (OuterVolumeSpecName: "kube-api-access-mhddw") pod "025427dc-d729-4b53-b665-de357c409d97" (UID: "025427dc-d729-4b53-b665-de357c409d97"). InnerVolumeSpecName "kube-api-access-mhddw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:07:20 crc kubenswrapper[4828]: I1210 19:07:20.221496 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/025427dc-d729-4b53-b665-de357c409d97-util" (OuterVolumeSpecName: "util") pod "025427dc-d729-4b53-b665-de357c409d97" (UID: "025427dc-d729-4b53-b665-de357c409d97"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:07:20 crc kubenswrapper[4828]: I1210 19:07:20.308020 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhddw\" (UniqueName: \"kubernetes.io/projected/025427dc-d729-4b53-b665-de357c409d97-kube-api-access-mhddw\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:20 crc kubenswrapper[4828]: I1210 19:07:20.308282 4828 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/025427dc-d729-4b53-b665-de357c409d97-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:20 crc kubenswrapper[4828]: I1210 19:07:20.308405 4828 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/025427dc-d729-4b53-b665-de357c409d97-util\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:20 crc kubenswrapper[4828]: I1210 19:07:20.772434 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" event={"ID":"025427dc-d729-4b53-b665-de357c409d97","Type":"ContainerDied","Data":"56e7241d385a7cd8acd4b0cee4819b4fbf4223ee72b01142f709bb05077ddba1"} Dec 10 19:07:20 crc kubenswrapper[4828]: I1210 19:07:20.772476 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56e7241d385a7cd8acd4b0cee4819b4fbf4223ee72b01142f709bb05077ddba1" Dec 10 19:07:20 crc kubenswrapper[4828]: I1210 19:07:20.772530 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f" Dec 10 19:07:21 crc kubenswrapper[4828]: I1210 19:07:21.230237 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:07:21 crc kubenswrapper[4828]: I1210 19:07:21.230309 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.349605 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8f42x"] Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.350067 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovn-controller" containerID="cri-o://f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908" gracePeriod=30 Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.350139 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="northd" containerID="cri-o://279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e" gracePeriod=30 Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.350210 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="kube-rbac-proxy-node" containerID="cri-o://a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77" gracePeriod=30 Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.350264 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovn-acl-logging" containerID="cri-o://0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7" gracePeriod=30 Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.350197 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275" gracePeriod=30 Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.350355 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="sbdb" containerID="cri-o://db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba" gracePeriod=30 Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.350368 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="nbdb" containerID="cri-o://62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2" gracePeriod=30 Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.434397 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovnkube-controller" containerID="cri-o://a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867" gracePeriod=30 Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.785789 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovnkube-controller/3.log" Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.788318 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovn-acl-logging/0.log" Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.788737 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovn-controller/0.log" Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.789198 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerID="a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867" exitCode=0 Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.789227 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerID="db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba" exitCode=0 Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.789235 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerID="62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2" exitCode=0 Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.789246 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerID="279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e" exitCode=0 Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.789253 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerID="0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7" exitCode=143 Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.789261 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerID="f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908" exitCode=143 Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.789253 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerDied","Data":"a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867"} Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.789297 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerDied","Data":"db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba"} Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.789311 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerDied","Data":"62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2"} Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.789321 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerDied","Data":"279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e"} Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.789329 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerDied","Data":"0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7"} Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.789338 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerDied","Data":"f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908"} Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.789353 4828 scope.go:117] "RemoveContainer" containerID="7e0ef04a88803858aabdbc8c1b702a19cc044e5df4e33e4d1a719f4d506e7c82" Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.791581 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xdxgd_c6e55203-65ae-4767-9a96-f309bc41ddf3/kube-multus/2.log" Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.792054 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xdxgd_c6e55203-65ae-4767-9a96-f309bc41ddf3/kube-multus/1.log" Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.792089 4828 generic.go:334] "Generic (PLEG): container finished" podID="c6e55203-65ae-4767-9a96-f309bc41ddf3" containerID="ea927775625b0de792d093bd00de1d34b82f3826d19fa5ddaf39be9b182f20ab" exitCode=2 Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.792110 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xdxgd" event={"ID":"c6e55203-65ae-4767-9a96-f309bc41ddf3","Type":"ContainerDied","Data":"ea927775625b0de792d093bd00de1d34b82f3826d19fa5ddaf39be9b182f20ab"} Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.792444 4828 scope.go:117] "RemoveContainer" containerID="ea927775625b0de792d093bd00de1d34b82f3826d19fa5ddaf39be9b182f20ab" Dec 10 19:07:22 crc kubenswrapper[4828]: E1210 19:07:22.792590 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-xdxgd_openshift-multus(c6e55203-65ae-4767-9a96-f309bc41ddf3)\"" pod="openshift-multus/multus-xdxgd" podUID="c6e55203-65ae-4767-9a96-f309bc41ddf3" Dec 10 19:07:22 crc kubenswrapper[4828]: I1210 19:07:22.823157 4828 scope.go:117] "RemoveContainer" containerID="bbe415208b7c4a043d3a9466d81931a16b2d600a3b5605eb89c7d2c2ffb1b71b" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.632743 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovn-acl-logging/0.log" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.633518 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovn-controller/0.log" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.633939 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.681776 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-vwlbb"] Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.682042 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="025427dc-d729-4b53-b665-de357c409d97" containerName="pull" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682055 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="025427dc-d729-4b53-b665-de357c409d97" containerName="pull" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.682068 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="kubecfg-setup" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682075 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="kubecfg-setup" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.682080 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovnkube-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682086 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovnkube-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.682093 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovnkube-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682099 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovnkube-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.682106 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="northd" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682112 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="northd" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.682124 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="nbdb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682130 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="nbdb" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.682137 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovn-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682142 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovn-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.682149 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="025427dc-d729-4b53-b665-de357c409d97" containerName="util" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682155 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="025427dc-d729-4b53-b665-de357c409d97" containerName="util" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.682162 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="sbdb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682168 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="sbdb" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.682177 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="025427dc-d729-4b53-b665-de357c409d97" containerName="extract" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682183 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="025427dc-d729-4b53-b665-de357c409d97" containerName="extract" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.682189 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="kube-rbac-proxy-node" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682195 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="kube-rbac-proxy-node" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.682206 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovn-acl-logging" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682213 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovn-acl-logging" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.682220 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovnkube-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682226 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovnkube-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.682235 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="kube-rbac-proxy-ovn-metrics" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682240 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="kube-rbac-proxy-ovn-metrics" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682330 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovnkube-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682339 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovn-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682346 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovnkube-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682353 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovn-acl-logging" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682359 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="nbdb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682368 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="025427dc-d729-4b53-b665-de357c409d97" containerName="extract" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682374 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="kube-rbac-proxy-ovn-metrics" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682381 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="northd" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682391 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovnkube-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682398 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="kube-rbac-proxy-node" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682406 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="sbdb" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.682499 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovnkube-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682506 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovnkube-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.682516 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovnkube-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682521 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovnkube-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682606 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovnkube-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.682613 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerName="ovnkube-controller" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.684217 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758559 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-slash\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758612 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-run-ovn-kubernetes\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758630 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-cni-bin\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758652 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-env-overrides\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758680 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hdv4\" (UniqueName: \"kubernetes.io/projected/1f37f60f-fe12-4530-8ff1-ea3794c65597-kube-api-access-4hdv4\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758695 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-node-log\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758711 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-cni-netd\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758762 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-kubelet\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758797 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-systemd-units\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758815 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-slash" (OuterVolumeSpecName: "host-slash") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758826 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-log-socket\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758848 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-log-socket" (OuterVolumeSpecName: "log-socket") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758871 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758875 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-openvswitch\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758891 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758903 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovnkube-config\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758926 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-systemd\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758948 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-var-lib-cni-networks-ovn-kubernetes\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758958 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758973 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-etc-openvswitch\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758990 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-node-log" (OuterVolumeSpecName: "node-log") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.758997 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-run-netns\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.759018 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.759031 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovn-node-metrics-cert\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.759118 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.759152 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.759144 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.759164 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.759136 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.759329 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.759396 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.759426 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.759209 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-var-lib-openvswitch\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.759775 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovnkube-script-lib\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.759824 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-ovn\") pod \"1f37f60f-fe12-4530-8ff1-ea3794c65597\" (UID: \"1f37f60f-fe12-4530-8ff1-ea3794c65597\") " Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.759963 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-cni-bin\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.759983 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-run-openvswitch\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760000 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-slash\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760026 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-run-systemd\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760052 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-systemd-units\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760072 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760099 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-node-log\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760118 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-log-socket\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760144 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-etc-openvswitch\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760177 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxbn7\" (UniqueName: \"kubernetes.io/projected/41eb203a-d66b-48b3-a049-438ad75bbc4a-kube-api-access-xxbn7\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760198 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-cni-netd\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760213 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-run-ovn-kubernetes\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760234 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-run-netns\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760248 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41eb203a-d66b-48b3-a049-438ad75bbc4a-ovn-node-metrics-cert\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760266 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41eb203a-d66b-48b3-a049-438ad75bbc4a-env-overrides\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760285 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-var-lib-openvswitch\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760303 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41eb203a-d66b-48b3-a049-438ad75bbc4a-ovnkube-script-lib\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760312 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760324 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-kubelet\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760360 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760394 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-run-ovn\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760426 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41eb203a-d66b-48b3-a049-438ad75bbc4a-ovnkube-config\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760565 4828 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760581 4828 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760595 4828 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760609 4828 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-slash\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760622 4828 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760636 4828 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760648 4828 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760660 4828 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-node-log\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760672 4828 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760684 4828 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760697 4828 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760708 4828 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-log-socket\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760719 4828 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760731 4828 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760743 4828 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760755 4828 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.760766 4828 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.771097 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.776413 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f37f60f-fe12-4530-8ff1-ea3794c65597-kube-api-access-4hdv4" (OuterVolumeSpecName: "kube-api-access-4hdv4") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "kube-api-access-4hdv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.783714 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "1f37f60f-fe12-4530-8ff1-ea3794c65597" (UID: "1f37f60f-fe12-4530-8ff1-ea3794c65597"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.798850 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xdxgd_c6e55203-65ae-4767-9a96-f309bc41ddf3/kube-multus/2.log" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.802565 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovn-acl-logging/0.log" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.803025 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8f42x_1f37f60f-fe12-4530-8ff1-ea3794c65597/ovn-controller/0.log" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.803350 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerID="6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275" exitCode=0 Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.803376 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f37f60f-fe12-4530-8ff1-ea3794c65597" containerID="a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77" exitCode=0 Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.803405 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerDied","Data":"6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275"} Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.803429 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.803446 4828 scope.go:117] "RemoveContainer" containerID="a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.803435 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerDied","Data":"a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77"} Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.803549 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8f42x" event={"ID":"1f37f60f-fe12-4530-8ff1-ea3794c65597","Type":"ContainerDied","Data":"d9089e51b51e3ffe9659d4b0684eb789eb5eeaa74d610fc958e741c6c6637568"} Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.828110 4828 scope.go:117] "RemoveContainer" containerID="db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.841456 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8f42x"] Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.845489 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8f42x"] Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.847621 4828 scope.go:117] "RemoveContainer" containerID="62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.859771 4828 scope.go:117] "RemoveContainer" containerID="279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.861491 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxbn7\" (UniqueName: \"kubernetes.io/projected/41eb203a-d66b-48b3-a049-438ad75bbc4a-kube-api-access-xxbn7\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.861531 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-cni-netd\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.861552 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-run-ovn-kubernetes\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.861587 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-run-netns\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.861612 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41eb203a-d66b-48b3-a049-438ad75bbc4a-ovn-node-metrics-cert\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.861638 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-cni-netd\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.861641 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41eb203a-d66b-48b3-a049-438ad75bbc4a-env-overrides\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.861737 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-var-lib-openvswitch\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.861769 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41eb203a-d66b-48b3-a049-438ad75bbc4a-ovnkube-script-lib\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.861832 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-var-lib-openvswitch\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.861683 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-run-netns\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.861681 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-run-ovn-kubernetes\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.861999 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-kubelet\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862040 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-run-ovn\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862062 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41eb203a-d66b-48b3-a049-438ad75bbc4a-ovnkube-config\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862085 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-cni-bin\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862110 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-run-openvswitch\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862126 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-run-ovn\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862132 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-slash\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862161 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-slash\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862190 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-kubelet\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862201 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-run-systemd\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862214 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-cni-bin\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862227 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-systemd-units\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862252 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862281 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-node-log\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862301 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-log-socket\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862330 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-etc-openvswitch\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862385 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hdv4\" (UniqueName: \"kubernetes.io/projected/1f37f60f-fe12-4530-8ff1-ea3794c65597-kube-api-access-4hdv4\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862404 4828 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1f37f60f-fe12-4530-8ff1-ea3794c65597-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862418 4828 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1f37f60f-fe12-4530-8ff1-ea3794c65597-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862449 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-etc-openvswitch\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862445 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/41eb203a-d66b-48b3-a049-438ad75bbc4a-env-overrides\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862493 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862500 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-systemd-units\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862527 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-node-log\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862528 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-log-socket\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862550 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-run-openvswitch\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862579 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/41eb203a-d66b-48b3-a049-438ad75bbc4a-run-systemd\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862579 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/41eb203a-d66b-48b3-a049-438ad75bbc4a-ovnkube-script-lib\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.862625 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/41eb203a-d66b-48b3-a049-438ad75bbc4a-ovnkube-config\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.865290 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/41eb203a-d66b-48b3-a049-438ad75bbc4a-ovn-node-metrics-cert\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.877227 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxbn7\" (UniqueName: \"kubernetes.io/projected/41eb203a-d66b-48b3-a049-438ad75bbc4a-kube-api-access-xxbn7\") pod \"ovnkube-node-vwlbb\" (UID: \"41eb203a-d66b-48b3-a049-438ad75bbc4a\") " pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.892959 4828 scope.go:117] "RemoveContainer" containerID="6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.905129 4828 scope.go:117] "RemoveContainer" containerID="a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.917600 4828 scope.go:117] "RemoveContainer" containerID="0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.933300 4828 scope.go:117] "RemoveContainer" containerID="f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.945798 4828 scope.go:117] "RemoveContainer" containerID="3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.965099 4828 scope.go:117] "RemoveContainer" containerID="a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.965622 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867\": container with ID starting with a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867 not found: ID does not exist" containerID="a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.965670 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867"} err="failed to get container status \"a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867\": rpc error: code = NotFound desc = could not find container \"a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867\": container with ID starting with a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867 not found: ID does not exist" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.965696 4828 scope.go:117] "RemoveContainer" containerID="db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.966134 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\": container with ID starting with db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba not found: ID does not exist" containerID="db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.966171 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba"} err="failed to get container status \"db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\": rpc error: code = NotFound desc = could not find container \"db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\": container with ID starting with db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba not found: ID does not exist" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.966202 4828 scope.go:117] "RemoveContainer" containerID="62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.966519 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\": container with ID starting with 62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2 not found: ID does not exist" containerID="62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.966561 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2"} err="failed to get container status \"62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\": rpc error: code = NotFound desc = could not find container \"62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\": container with ID starting with 62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2 not found: ID does not exist" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.966576 4828 scope.go:117] "RemoveContainer" containerID="279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.966853 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\": container with ID starting with 279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e not found: ID does not exist" containerID="279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.966875 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e"} err="failed to get container status \"279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\": rpc error: code = NotFound desc = could not find container \"279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\": container with ID starting with 279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e not found: ID does not exist" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.966925 4828 scope.go:117] "RemoveContainer" containerID="6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.967215 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\": container with ID starting with 6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275 not found: ID does not exist" containerID="6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.967268 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275"} err="failed to get container status \"6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\": rpc error: code = NotFound desc = could not find container \"6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\": container with ID starting with 6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275 not found: ID does not exist" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.967288 4828 scope.go:117] "RemoveContainer" containerID="a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.967530 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\": container with ID starting with a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77 not found: ID does not exist" containerID="a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.967560 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77"} err="failed to get container status \"a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\": rpc error: code = NotFound desc = could not find container \"a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\": container with ID starting with a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77 not found: ID does not exist" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.967580 4828 scope.go:117] "RemoveContainer" containerID="0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.967866 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\": container with ID starting with 0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7 not found: ID does not exist" containerID="0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.967897 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7"} err="failed to get container status \"0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\": rpc error: code = NotFound desc = could not find container \"0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\": container with ID starting with 0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7 not found: ID does not exist" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.967914 4828 scope.go:117] "RemoveContainer" containerID="f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.968187 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\": container with ID starting with f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908 not found: ID does not exist" containerID="f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.968225 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908"} err="failed to get container status \"f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\": rpc error: code = NotFound desc = could not find container \"f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\": container with ID starting with f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908 not found: ID does not exist" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.968252 4828 scope.go:117] "RemoveContainer" containerID="3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e" Dec 10 19:07:23 crc kubenswrapper[4828]: E1210 19:07:23.968548 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\": container with ID starting with 3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e not found: ID does not exist" containerID="3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.968576 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e"} err="failed to get container status \"3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\": rpc error: code = NotFound desc = could not find container \"3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\": container with ID starting with 3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e not found: ID does not exist" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.968593 4828 scope.go:117] "RemoveContainer" containerID="a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.968956 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867"} err="failed to get container status \"a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867\": rpc error: code = NotFound desc = could not find container \"a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867\": container with ID starting with a403c54c315fd96dabd4f0d8a67073acd288d4bdc25fab965845e34564b0f867 not found: ID does not exist" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.968978 4828 scope.go:117] "RemoveContainer" containerID="db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.969274 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba"} err="failed to get container status \"db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\": rpc error: code = NotFound desc = could not find container \"db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba\": container with ID starting with db038fda8c8b638185ded6c9d76acf3d0ef197866dc6978419f225354698ebba not found: ID does not exist" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.969297 4828 scope.go:117] "RemoveContainer" containerID="62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.969697 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2"} err="failed to get container status \"62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\": rpc error: code = NotFound desc = could not find container \"62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2\": container with ID starting with 62b3290c799c54de07ef2c8220320c20b00ecc28cdab29ae91278352045785f2 not found: ID does not exist" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.969719 4828 scope.go:117] "RemoveContainer" containerID="279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.970018 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e"} err="failed to get container status \"279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\": rpc error: code = NotFound desc = could not find container \"279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e\": container with ID starting with 279a352566a30cf257e35648a65904c2046ddcbcd5cefdfb5919fb1c8a46233e not found: ID does not exist" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.970044 4828 scope.go:117] "RemoveContainer" containerID="6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.970314 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275"} err="failed to get container status \"6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\": rpc error: code = NotFound desc = could not find container \"6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275\": container with ID starting with 6ade6ca6d2ea268f55770a7af320249bae0dd8ad811824015ff1d4898a9a0275 not found: ID does not exist" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.970336 4828 scope.go:117] "RemoveContainer" containerID="a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.970669 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77"} err="failed to get container status \"a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\": rpc error: code = NotFound desc = could not find container \"a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77\": container with ID starting with a9e10a3e8b9ec2980ff5a9098f9fc95b9d7fc1ba91d624b4f243c7f7a785dd77 not found: ID does not exist" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.970687 4828 scope.go:117] "RemoveContainer" containerID="0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.971040 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7"} err="failed to get container status \"0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\": rpc error: code = NotFound desc = could not find container \"0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7\": container with ID starting with 0ff3be4ffaebdb1da3b35293da8c73af55b43098614e1274c863dee9deff62f7 not found: ID does not exist" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.971069 4828 scope.go:117] "RemoveContainer" containerID="f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.971392 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908"} err="failed to get container status \"f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\": rpc error: code = NotFound desc = could not find container \"f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908\": container with ID starting with f6d09fe0825ded1eac2745908032239daf86f60adee669955aed33509af48908 not found: ID does not exist" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.971423 4828 scope.go:117] "RemoveContainer" containerID="3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e" Dec 10 19:07:23 crc kubenswrapper[4828]: I1210 19:07:23.971684 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e"} err="failed to get container status \"3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\": rpc error: code = NotFound desc = could not find container \"3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e\": container with ID starting with 3dd20e74fa3afd14b4c743ff80021687a692d98bf376109f3033851de3d95a7e not found: ID does not exist" Dec 10 19:07:24 crc kubenswrapper[4828]: I1210 19:07:24.003657 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:24 crc kubenswrapper[4828]: I1210 19:07:24.810514 4828 generic.go:334] "Generic (PLEG): container finished" podID="41eb203a-d66b-48b3-a049-438ad75bbc4a" containerID="2df19c18fc96baaa80626f9863ba8c7928ea63eb3601a08c958aa74f0b9e9b72" exitCode=0 Dec 10 19:07:24 crc kubenswrapper[4828]: I1210 19:07:24.810618 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" event={"ID":"41eb203a-d66b-48b3-a049-438ad75bbc4a","Type":"ContainerDied","Data":"2df19c18fc96baaa80626f9863ba8c7928ea63eb3601a08c958aa74f0b9e9b72"} Dec 10 19:07:24 crc kubenswrapper[4828]: I1210 19:07:24.811027 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" event={"ID":"41eb203a-d66b-48b3-a049-438ad75bbc4a","Type":"ContainerStarted","Data":"8dd44415e5a54e72ed2e82ccaaa48b6f8e7f128d375f6645ac864bcf9d2f0dc1"} Dec 10 19:07:25 crc kubenswrapper[4828]: I1210 19:07:25.796839 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f37f60f-fe12-4530-8ff1-ea3794c65597" path="/var/lib/kubelet/pods/1f37f60f-fe12-4530-8ff1-ea3794c65597/volumes" Dec 10 19:07:25 crc kubenswrapper[4828]: I1210 19:07:25.821011 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" event={"ID":"41eb203a-d66b-48b3-a049-438ad75bbc4a","Type":"ContainerStarted","Data":"b2065f9d287b5576dc7544cb75af84075cc9ef0fa0550f50516b5082ca500c0a"} Dec 10 19:07:25 crc kubenswrapper[4828]: I1210 19:07:25.821057 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" event={"ID":"41eb203a-d66b-48b3-a049-438ad75bbc4a","Type":"ContainerStarted","Data":"f60dbfcb598d6679591bf83429545547419709421b3b4e528327b12d39901736"} Dec 10 19:07:25 crc kubenswrapper[4828]: I1210 19:07:25.821070 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" event={"ID":"41eb203a-d66b-48b3-a049-438ad75bbc4a","Type":"ContainerStarted","Data":"ca44ba75fb9e961b087830399c8c95ff35f984d6831a60b6dcfab3dde316bd44"} Dec 10 19:07:25 crc kubenswrapper[4828]: I1210 19:07:25.821082 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" event={"ID":"41eb203a-d66b-48b3-a049-438ad75bbc4a","Type":"ContainerStarted","Data":"b6aea7c3616d24ef71600e651eaf118c901f80eb75ca05e22712160d28fa5ec3"} Dec 10 19:07:25 crc kubenswrapper[4828]: I1210 19:07:25.821093 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" event={"ID":"41eb203a-d66b-48b3-a049-438ad75bbc4a","Type":"ContainerStarted","Data":"eb903dd9e2f34a389995ca89100121ef153f83f590ffff727b317f532eca1935"} Dec 10 19:07:25 crc kubenswrapper[4828]: I1210 19:07:25.821105 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" event={"ID":"41eb203a-d66b-48b3-a049-438ad75bbc4a","Type":"ContainerStarted","Data":"948d264d782c06a41f70a45c5633c37fc09d2040914c5588b1d2ad6bfb4e8148"} Dec 10 19:07:27 crc kubenswrapper[4828]: I1210 19:07:27.845268 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" event={"ID":"41eb203a-d66b-48b3-a049-438ad75bbc4a","Type":"ContainerStarted","Data":"41a21edb7937b60f73714d648d37513a24271db683cd591275cfde61df52fa87"} Dec 10 19:07:30 crc kubenswrapper[4828]: I1210 19:07:30.864614 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" event={"ID":"41eb203a-d66b-48b3-a049-438ad75bbc4a","Type":"ContainerStarted","Data":"11ad5a3c610e8631876d0e3d8f7c55948f8ca6357f2802910ccfaeea554aa831"} Dec 10 19:07:30 crc kubenswrapper[4828]: I1210 19:07:30.865404 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:30 crc kubenswrapper[4828]: I1210 19:07:30.865417 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:30 crc kubenswrapper[4828]: I1210 19:07:30.926159 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:30 crc kubenswrapper[4828]: I1210 19:07:30.942077 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:30 crc kubenswrapper[4828]: I1210 19:07:30.976863 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" podStartSLOduration=7.976841809 podStartE2EDuration="7.976841809s" podCreationTimestamp="2025-12-10 19:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:07:30.919981951 +0000 UTC m=+731.430592956" watchObservedRunningTime="2025-12-10 19:07:30.976841809 +0000 UTC m=+731.487452814" Dec 10 19:07:31 crc kubenswrapper[4828]: I1210 19:07:31.871158 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.417183 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk"] Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.417879 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.424347 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-vk6m4" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.424571 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.424721 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.437687 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk"] Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.495448 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqgtf\" (UniqueName: \"kubernetes.io/projected/fe95ea54-da09-4ec9-a930-5be068b09954-kube-api-access-sqgtf\") pod \"obo-prometheus-operator-668cf9dfbb-vpjkk\" (UID: \"fe95ea54-da09-4ec9-a930-5be068b09954\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.535383 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf"] Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.536140 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.539534 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-jcphw" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.539599 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.549461 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz"] Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.550194 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.557917 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf"] Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.569708 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz"] Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.597128 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqgtf\" (UniqueName: \"kubernetes.io/projected/fe95ea54-da09-4ec9-a930-5be068b09954-kube-api-access-sqgtf\") pod \"obo-prometheus-operator-668cf9dfbb-vpjkk\" (UID: \"fe95ea54-da09-4ec9-a930-5be068b09954\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.616384 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqgtf\" (UniqueName: \"kubernetes.io/projected/fe95ea54-da09-4ec9-a930-5be068b09954-kube-api-access-sqgtf\") pod \"obo-prometheus-operator-668cf9dfbb-vpjkk\" (UID: \"fe95ea54-da09-4ec9-a930-5be068b09954\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.650508 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-flqjw"] Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.651240 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.652784 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-cb6pq" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.652784 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.698528 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/68b3a59c-1ec9-4381-a92a-810b554d76e0-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6774fff8-2glrf\" (UID: \"68b3a59c-1ec9-4381-a92a-810b554d76e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.698627 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b5194787-06b3-40a3-92dd-7a7471ba5829-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6774fff8-sbnvz\" (UID: \"b5194787-06b3-40a3-92dd-7a7471ba5829\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.698657 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/68b3a59c-1ec9-4381-a92a-810b554d76e0-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6774fff8-2glrf\" (UID: \"68b3a59c-1ec9-4381-a92a-810b554d76e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.698684 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b5194787-06b3-40a3-92dd-7a7471ba5829-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6774fff8-sbnvz\" (UID: \"b5194787-06b3-40a3-92dd-7a7471ba5829\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.711872 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-flqjw"] Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.743585 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.750240 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-57l2h"] Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.751170 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.753975 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-sb2cx" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.765380 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-57l2h"] Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.788641 4828 scope.go:117] "RemoveContainer" containerID="ea927775625b0de792d093bd00de1d34b82f3826d19fa5ddaf39be9b182f20ab" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.790940 4828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-vpjkk_openshift-operators_fe95ea54-da09-4ec9-a930-5be068b09954_0(9cf462b1f361674ae7c2a28e213d6ce69a2d699872aa01a4711e855b50e4db05): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.790995 4828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-vpjkk_openshift-operators_fe95ea54-da09-4ec9-a930-5be068b09954_0(9cf462b1f361674ae7c2a28e213d6ce69a2d699872aa01a4711e855b50e4db05): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.791014 4828 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-vpjkk_openshift-operators_fe95ea54-da09-4ec9-a930-5be068b09954_0(9cf462b1f361674ae7c2a28e213d6ce69a2d699872aa01a4711e855b50e4db05): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.791065 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-vpjkk_openshift-operators(fe95ea54-da09-4ec9-a930-5be068b09954)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-vpjkk_openshift-operators(fe95ea54-da09-4ec9-a930-5be068b09954)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-vpjkk_openshift-operators_fe95ea54-da09-4ec9-a930-5be068b09954_0(9cf462b1f361674ae7c2a28e213d6ce69a2d699872aa01a4711e855b50e4db05): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" podUID="fe95ea54-da09-4ec9-a930-5be068b09954" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.799413 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b5194787-06b3-40a3-92dd-7a7471ba5829-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6774fff8-sbnvz\" (UID: \"b5194787-06b3-40a3-92dd-7a7471ba5829\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.799460 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/68b3a59c-1ec9-4381-a92a-810b554d76e0-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6774fff8-2glrf\" (UID: \"68b3a59c-1ec9-4381-a92a-810b554d76e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.799492 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmgtc\" (UniqueName: \"kubernetes.io/projected/b55b0a41-ae3f-4acf-bf2d-0378868dca03-kube-api-access-dmgtc\") pod \"observability-operator-d8bb48f5d-flqjw\" (UID: \"b55b0a41-ae3f-4acf-bf2d-0378868dca03\") " pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.799552 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/b55b0a41-ae3f-4acf-bf2d-0378868dca03-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-flqjw\" (UID: \"b55b0a41-ae3f-4acf-bf2d-0378868dca03\") " pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.799620 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b5194787-06b3-40a3-92dd-7a7471ba5829-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6774fff8-sbnvz\" (UID: \"b5194787-06b3-40a3-92dd-7a7471ba5829\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.799661 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/68b3a59c-1ec9-4381-a92a-810b554d76e0-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6774fff8-2glrf\" (UID: \"68b3a59c-1ec9-4381-a92a-810b554d76e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.803513 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b5194787-06b3-40a3-92dd-7a7471ba5829-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6774fff8-sbnvz\" (UID: \"b5194787-06b3-40a3-92dd-7a7471ba5829\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.804401 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b5194787-06b3-40a3-92dd-7a7471ba5829-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6774fff8-sbnvz\" (UID: \"b5194787-06b3-40a3-92dd-7a7471ba5829\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.804781 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/68b3a59c-1ec9-4381-a92a-810b554d76e0-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-6774fff8-2glrf\" (UID: \"68b3a59c-1ec9-4381-a92a-810b554d76e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.807252 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/68b3a59c-1ec9-4381-a92a-810b554d76e0-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-6774fff8-2glrf\" (UID: \"68b3a59c-1ec9-4381-a92a-810b554d76e0\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.851205 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.863063 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.881557 4828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6774fff8-2glrf_openshift-operators_68b3a59c-1ec9-4381-a92a-810b554d76e0_0(2b3e4c6b517362bb16a84f4c9cd3418b7b8eb7c9a9815f6208404af6453e4818): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.881637 4828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6774fff8-2glrf_openshift-operators_68b3a59c-1ec9-4381-a92a-810b554d76e0_0(2b3e4c6b517362bb16a84f4c9cd3418b7b8eb7c9a9815f6208404af6453e4818): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.881666 4828 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6774fff8-2glrf_openshift-operators_68b3a59c-1ec9-4381-a92a-810b554d76e0_0(2b3e4c6b517362bb16a84f4c9cd3418b7b8eb7c9a9815f6208404af6453e4818): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.881732 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6774fff8-2glrf_openshift-operators(68b3a59c-1ec9-4381-a92a-810b554d76e0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6774fff8-2glrf_openshift-operators(68b3a59c-1ec9-4381-a92a-810b554d76e0)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6774fff8-2glrf_openshift-operators_68b3a59c-1ec9-4381-a92a-810b554d76e0_0(2b3e4c6b517362bb16a84f4c9cd3418b7b8eb7c9a9815f6208404af6453e4818): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" podUID="68b3a59c-1ec9-4381-a92a-810b554d76e0" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.883788 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.884452 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.900909 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmgtc\" (UniqueName: \"kubernetes.io/projected/b55b0a41-ae3f-4acf-bf2d-0378868dca03-kube-api-access-dmgtc\") pod \"observability-operator-d8bb48f5d-flqjw\" (UID: \"b55b0a41-ae3f-4acf-bf2d-0378868dca03\") " pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.900973 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxp8x\" (UniqueName: \"kubernetes.io/projected/ebb4036a-1bd8-4636-8bea-63b0a6b4cc38-kube-api-access-dxp8x\") pod \"perses-operator-5446b9c989-57l2h\" (UID: \"ebb4036a-1bd8-4636-8bea-63b0a6b4cc38\") " pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.901016 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/b55b0a41-ae3f-4acf-bf2d-0378868dca03-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-flqjw\" (UID: \"b55b0a41-ae3f-4acf-bf2d-0378868dca03\") " pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.901033 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/ebb4036a-1bd8-4636-8bea-63b0a6b4cc38-openshift-service-ca\") pod \"perses-operator-5446b9c989-57l2h\" (UID: \"ebb4036a-1bd8-4636-8bea-63b0a6b4cc38\") " pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.907080 4828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6774fff8-sbnvz_openshift-operators_b5194787-06b3-40a3-92dd-7a7471ba5829_0(cfc5031b78df2e22ae9820ff14df0a3440d412d86fe5fc35af16d2d55985c7ae): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.907132 4828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6774fff8-sbnvz_openshift-operators_b5194787-06b3-40a3-92dd-7a7471ba5829_0(cfc5031b78df2e22ae9820ff14df0a3440d412d86fe5fc35af16d2d55985c7ae): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.907154 4828 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6774fff8-sbnvz_openshift-operators_b5194787-06b3-40a3-92dd-7a7471ba5829_0(cfc5031b78df2e22ae9820ff14df0a3440d412d86fe5fc35af16d2d55985c7ae): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.907193 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6774fff8-sbnvz_openshift-operators(b5194787-06b3-40a3-92dd-7a7471ba5829)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6774fff8-sbnvz_openshift-operators(b5194787-06b3-40a3-92dd-7a7471ba5829)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6774fff8-sbnvz_openshift-operators_b5194787-06b3-40a3-92dd-7a7471ba5829_0(cfc5031b78df2e22ae9820ff14df0a3440d412d86fe5fc35af16d2d55985c7ae): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" podUID="b5194787-06b3-40a3-92dd-7a7471ba5829" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.908704 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/b55b0a41-ae3f-4acf-bf2d-0378868dca03-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-flqjw\" (UID: \"b55b0a41-ae3f-4acf-bf2d-0378868dca03\") " pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.918233 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmgtc\" (UniqueName: \"kubernetes.io/projected/b55b0a41-ae3f-4acf-bf2d-0378868dca03-kube-api-access-dmgtc\") pod \"observability-operator-d8bb48f5d-flqjw\" (UID: \"b55b0a41-ae3f-4acf-bf2d-0378868dca03\") " pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.934000 4828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-vpjkk_openshift-operators_fe95ea54-da09-4ec9-a930-5be068b09954_0(974ae4b28d00973b4520846faa0fafbc8d042c94841bafd2a262132234e7ebce): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.934067 4828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-vpjkk_openshift-operators_fe95ea54-da09-4ec9-a930-5be068b09954_0(974ae4b28d00973b4520846faa0fafbc8d042c94841bafd2a262132234e7ebce): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.934090 4828 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-vpjkk_openshift-operators_fe95ea54-da09-4ec9-a930-5be068b09954_0(974ae4b28d00973b4520846faa0fafbc8d042c94841bafd2a262132234e7ebce): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.934138 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-vpjkk_openshift-operators(fe95ea54-da09-4ec9-a930-5be068b09954)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-vpjkk_openshift-operators(fe95ea54-da09-4ec9-a930-5be068b09954)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-vpjkk_openshift-operators_fe95ea54-da09-4ec9-a930-5be068b09954_0(974ae4b28d00973b4520846faa0fafbc8d042c94841bafd2a262132234e7ebce): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" podUID="fe95ea54-da09-4ec9-a930-5be068b09954" Dec 10 19:07:33 crc kubenswrapper[4828]: I1210 19:07:33.969995 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.994756 4828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-flqjw_openshift-operators_b55b0a41-ae3f-4acf-bf2d-0378868dca03_0(d1a2db0a5e1759352886380c9f411ff2b30c99e60afe31dac60ca20e74fd8f6c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.994849 4828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-flqjw_openshift-operators_b55b0a41-ae3f-4acf-bf2d-0378868dca03_0(d1a2db0a5e1759352886380c9f411ff2b30c99e60afe31dac60ca20e74fd8f6c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.994877 4828 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-flqjw_openshift-operators_b55b0a41-ae3f-4acf-bf2d-0378868dca03_0(d1a2db0a5e1759352886380c9f411ff2b30c99e60afe31dac60ca20e74fd8f6c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:07:33 crc kubenswrapper[4828]: E1210 19:07:33.994921 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-flqjw_openshift-operators(b55b0a41-ae3f-4acf-bf2d-0378868dca03)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-flqjw_openshift-operators(b55b0a41-ae3f-4acf-bf2d-0378868dca03)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-flqjw_openshift-operators_b55b0a41-ae3f-4acf-bf2d-0378868dca03_0(d1a2db0a5e1759352886380c9f411ff2b30c99e60afe31dac60ca20e74fd8f6c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" podUID="b55b0a41-ae3f-4acf-bf2d-0378868dca03" Dec 10 19:07:34 crc kubenswrapper[4828]: I1210 19:07:34.004364 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxp8x\" (UniqueName: \"kubernetes.io/projected/ebb4036a-1bd8-4636-8bea-63b0a6b4cc38-kube-api-access-dxp8x\") pod \"perses-operator-5446b9c989-57l2h\" (UID: \"ebb4036a-1bd8-4636-8bea-63b0a6b4cc38\") " pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:07:34 crc kubenswrapper[4828]: I1210 19:07:34.004448 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/ebb4036a-1bd8-4636-8bea-63b0a6b4cc38-openshift-service-ca\") pod \"perses-operator-5446b9c989-57l2h\" (UID: \"ebb4036a-1bd8-4636-8bea-63b0a6b4cc38\") " pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:07:34 crc kubenswrapper[4828]: I1210 19:07:34.005854 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/ebb4036a-1bd8-4636-8bea-63b0a6b4cc38-openshift-service-ca\") pod \"perses-operator-5446b9c989-57l2h\" (UID: \"ebb4036a-1bd8-4636-8bea-63b0a6b4cc38\") " pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:07:34 crc kubenswrapper[4828]: I1210 19:07:34.050588 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxp8x\" (UniqueName: \"kubernetes.io/projected/ebb4036a-1bd8-4636-8bea-63b0a6b4cc38-kube-api-access-dxp8x\") pod \"perses-operator-5446b9c989-57l2h\" (UID: \"ebb4036a-1bd8-4636-8bea-63b0a6b4cc38\") " pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:07:34 crc kubenswrapper[4828]: I1210 19:07:34.073115 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:07:34 crc kubenswrapper[4828]: E1210 19:07:34.093422 4828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-57l2h_openshift-operators_ebb4036a-1bd8-4636-8bea-63b0a6b4cc38_0(e25a729ac26b3423c93ab7b6dbd2335708fd335f2041febd0ec22cd1d9402a61): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:34 crc kubenswrapper[4828]: E1210 19:07:34.093472 4828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-57l2h_openshift-operators_ebb4036a-1bd8-4636-8bea-63b0a6b4cc38_0(e25a729ac26b3423c93ab7b6dbd2335708fd335f2041febd0ec22cd1d9402a61): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:07:34 crc kubenswrapper[4828]: E1210 19:07:34.093493 4828 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-57l2h_openshift-operators_ebb4036a-1bd8-4636-8bea-63b0a6b4cc38_0(e25a729ac26b3423c93ab7b6dbd2335708fd335f2041febd0ec22cd1d9402a61): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:07:34 crc kubenswrapper[4828]: E1210 19:07:34.093541 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-57l2h_openshift-operators(ebb4036a-1bd8-4636-8bea-63b0a6b4cc38)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-57l2h_openshift-operators(ebb4036a-1bd8-4636-8bea-63b0a6b4cc38)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-57l2h_openshift-operators_ebb4036a-1bd8-4636-8bea-63b0a6b4cc38_0(e25a729ac26b3423c93ab7b6dbd2335708fd335f2041febd0ec22cd1d9402a61): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-57l2h" podUID="ebb4036a-1bd8-4636-8bea-63b0a6b4cc38" Dec 10 19:07:34 crc kubenswrapper[4828]: I1210 19:07:34.890430 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xdxgd_c6e55203-65ae-4767-9a96-f309bc41ddf3/kube-multus/2.log" Dec 10 19:07:34 crc kubenswrapper[4828]: I1210 19:07:34.890538 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" Dec 10 19:07:34 crc kubenswrapper[4828]: I1210 19:07:34.890531 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xdxgd" event={"ID":"c6e55203-65ae-4767-9a96-f309bc41ddf3","Type":"ContainerStarted","Data":"447398787db9e4ae8b31cba0ae8f726351e9257b3e407e90f884aef38ffdee59"} Dec 10 19:07:34 crc kubenswrapper[4828]: I1210 19:07:34.890552 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:07:34 crc kubenswrapper[4828]: I1210 19:07:34.890630 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:07:34 crc kubenswrapper[4828]: I1210 19:07:34.890734 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" Dec 10 19:07:34 crc kubenswrapper[4828]: I1210 19:07:34.891054 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" Dec 10 19:07:34 crc kubenswrapper[4828]: I1210 19:07:34.891269 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:07:34 crc kubenswrapper[4828]: I1210 19:07:34.891305 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" Dec 10 19:07:34 crc kubenswrapper[4828]: I1210 19:07:34.891449 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:07:34 crc kubenswrapper[4828]: E1210 19:07:34.979543 4828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-flqjw_openshift-operators_b55b0a41-ae3f-4acf-bf2d-0378868dca03_0(8916fcc3ecba85161ea32647cb64ed198d0f0c4d1f5e5a7f10f988b3209ac71d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:34 crc kubenswrapper[4828]: E1210 19:07:34.979895 4828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-flqjw_openshift-operators_b55b0a41-ae3f-4acf-bf2d-0378868dca03_0(8916fcc3ecba85161ea32647cb64ed198d0f0c4d1f5e5a7f10f988b3209ac71d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:07:34 crc kubenswrapper[4828]: E1210 19:07:34.979916 4828 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-flqjw_openshift-operators_b55b0a41-ae3f-4acf-bf2d-0378868dca03_0(8916fcc3ecba85161ea32647cb64ed198d0f0c4d1f5e5a7f10f988b3209ac71d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:07:34 crc kubenswrapper[4828]: E1210 19:07:34.979954 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-flqjw_openshift-operators(b55b0a41-ae3f-4acf-bf2d-0378868dca03)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-flqjw_openshift-operators(b55b0a41-ae3f-4acf-bf2d-0378868dca03)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-flqjw_openshift-operators_b55b0a41-ae3f-4acf-bf2d-0378868dca03_0(8916fcc3ecba85161ea32647cb64ed198d0f0c4d1f5e5a7f10f988b3209ac71d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" podUID="b55b0a41-ae3f-4acf-bf2d-0378868dca03" Dec 10 19:07:34 crc kubenswrapper[4828]: E1210 19:07:34.989713 4828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6774fff8-2glrf_openshift-operators_68b3a59c-1ec9-4381-a92a-810b554d76e0_0(ce4375a71728660a7e8d873bad88eea417e793551a9b6c43656b1a692d856668): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:34 crc kubenswrapper[4828]: E1210 19:07:34.989884 4828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6774fff8-2glrf_openshift-operators_68b3a59c-1ec9-4381-a92a-810b554d76e0_0(ce4375a71728660a7e8d873bad88eea417e793551a9b6c43656b1a692d856668): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" Dec 10 19:07:34 crc kubenswrapper[4828]: E1210 19:07:34.989914 4828 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6774fff8-2glrf_openshift-operators_68b3a59c-1ec9-4381-a92a-810b554d76e0_0(ce4375a71728660a7e8d873bad88eea417e793551a9b6c43656b1a692d856668): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" Dec 10 19:07:34 crc kubenswrapper[4828]: E1210 19:07:34.989974 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6774fff8-2glrf_openshift-operators(68b3a59c-1ec9-4381-a92a-810b554d76e0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6774fff8-2glrf_openshift-operators(68b3a59c-1ec9-4381-a92a-810b554d76e0)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6774fff8-2glrf_openshift-operators_68b3a59c-1ec9-4381-a92a-810b554d76e0_0(ce4375a71728660a7e8d873bad88eea417e793551a9b6c43656b1a692d856668): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" podUID="68b3a59c-1ec9-4381-a92a-810b554d76e0" Dec 10 19:07:34 crc kubenswrapper[4828]: E1210 19:07:34.995923 4828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-57l2h_openshift-operators_ebb4036a-1bd8-4636-8bea-63b0a6b4cc38_0(f5fc1c51339ad5da39e9825d9e88df1cfa104253a369a15bdbb0e04f563ed80d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:34 crc kubenswrapper[4828]: E1210 19:07:34.995988 4828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-57l2h_openshift-operators_ebb4036a-1bd8-4636-8bea-63b0a6b4cc38_0(f5fc1c51339ad5da39e9825d9e88df1cfa104253a369a15bdbb0e04f563ed80d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:07:34 crc kubenswrapper[4828]: E1210 19:07:34.996007 4828 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-57l2h_openshift-operators_ebb4036a-1bd8-4636-8bea-63b0a6b4cc38_0(f5fc1c51339ad5da39e9825d9e88df1cfa104253a369a15bdbb0e04f563ed80d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:07:34 crc kubenswrapper[4828]: E1210 19:07:34.996056 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-57l2h_openshift-operators(ebb4036a-1bd8-4636-8bea-63b0a6b4cc38)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-57l2h_openshift-operators(ebb4036a-1bd8-4636-8bea-63b0a6b4cc38)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-57l2h_openshift-operators_ebb4036a-1bd8-4636-8bea-63b0a6b4cc38_0(f5fc1c51339ad5da39e9825d9e88df1cfa104253a369a15bdbb0e04f563ed80d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-57l2h" podUID="ebb4036a-1bd8-4636-8bea-63b0a6b4cc38" Dec 10 19:07:35 crc kubenswrapper[4828]: E1210 19:07:35.015365 4828 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6774fff8-sbnvz_openshift-operators_b5194787-06b3-40a3-92dd-7a7471ba5829_0(3dce258cda9a0da0379021849cb8b63db98a93e9f573f37f241040284f965c63): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 10 19:07:35 crc kubenswrapper[4828]: E1210 19:07:35.015435 4828 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6774fff8-sbnvz_openshift-operators_b5194787-06b3-40a3-92dd-7a7471ba5829_0(3dce258cda9a0da0379021849cb8b63db98a93e9f573f37f241040284f965c63): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" Dec 10 19:07:35 crc kubenswrapper[4828]: E1210 19:07:35.015468 4828 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6774fff8-sbnvz_openshift-operators_b5194787-06b3-40a3-92dd-7a7471ba5829_0(3dce258cda9a0da0379021849cb8b63db98a93e9f573f37f241040284f965c63): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" Dec 10 19:07:35 crc kubenswrapper[4828]: E1210 19:07:35.015531 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-6774fff8-sbnvz_openshift-operators(b5194787-06b3-40a3-92dd-7a7471ba5829)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-6774fff8-sbnvz_openshift-operators(b5194787-06b3-40a3-92dd-7a7471ba5829)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-6774fff8-sbnvz_openshift-operators_b5194787-06b3-40a3-92dd-7a7471ba5829_0(3dce258cda9a0da0379021849cb8b63db98a93e9f573f37f241040284f965c63): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" podUID="b5194787-06b3-40a3-92dd-7a7471ba5829" Dec 10 19:07:46 crc kubenswrapper[4828]: I1210 19:07:46.788012 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" Dec 10 19:07:46 crc kubenswrapper[4828]: I1210 19:07:46.788978 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" Dec 10 19:07:47 crc kubenswrapper[4828]: I1210 19:07:47.096490 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf"] Dec 10 19:07:47 crc kubenswrapper[4828]: I1210 19:07:47.788682 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:07:47 crc kubenswrapper[4828]: I1210 19:07:47.788766 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" Dec 10 19:07:47 crc kubenswrapper[4828]: I1210 19:07:47.789339 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:07:47 crc kubenswrapper[4828]: I1210 19:07:47.789705 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" Dec 10 19:07:47 crc kubenswrapper[4828]: I1210 19:07:47.974732 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" event={"ID":"68b3a59c-1ec9-4381-a92a-810b554d76e0","Type":"ContainerStarted","Data":"31d8bb3893604124aae7d03caa6a7c36f2efa2882aeea37c47cecc26ce3825c9"} Dec 10 19:07:48 crc kubenswrapper[4828]: I1210 19:07:48.281221 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz"] Dec 10 19:07:48 crc kubenswrapper[4828]: I1210 19:07:48.375098 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-flqjw"] Dec 10 19:07:48 crc kubenswrapper[4828]: I1210 19:07:48.787953 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" Dec 10 19:07:48 crc kubenswrapper[4828]: I1210 19:07:48.788687 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" Dec 10 19:07:48 crc kubenswrapper[4828]: I1210 19:07:48.981332 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" event={"ID":"b5194787-06b3-40a3-92dd-7a7471ba5829","Type":"ContainerStarted","Data":"29ecd6de1f95dcf23a89e991c866b3ca9b8d884e59a62cf3d10c7b7e9170b92b"} Dec 10 19:07:49 crc kubenswrapper[4828]: I1210 19:07:49.754039 4828 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 10 19:07:49 crc kubenswrapper[4828]: I1210 19:07:49.990505 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" event={"ID":"b55b0a41-ae3f-4acf-bf2d-0378868dca03","Type":"ContainerStarted","Data":"249030d1738fc6ec6837ed73831a05bdbe772ee053a23ef2fac845995e50b479"} Dec 10 19:07:50 crc kubenswrapper[4828]: I1210 19:07:50.209655 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk"] Dec 10 19:07:50 crc kubenswrapper[4828]: I1210 19:07:50.788378 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:07:50 crc kubenswrapper[4828]: I1210 19:07:50.788859 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:07:51 crc kubenswrapper[4828]: I1210 19:07:51.018501 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" event={"ID":"fe95ea54-da09-4ec9-a930-5be068b09954","Type":"ContainerStarted","Data":"6df9f8153a1730175417d1d74ee97a16b32abac18055af06594e960972400a3a"} Dec 10 19:07:51 crc kubenswrapper[4828]: I1210 19:07:51.184482 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-57l2h"] Dec 10 19:07:51 crc kubenswrapper[4828]: W1210 19:07:51.195674 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebb4036a_1bd8_4636_8bea_63b0a6b4cc38.slice/crio-736b8e5a13f385c1bbac3359e8a86c3f58632751d9b57af133cd6434b5c9e33a WatchSource:0}: Error finding container 736b8e5a13f385c1bbac3359e8a86c3f58632751d9b57af133cd6434b5c9e33a: Status 404 returned error can't find the container with id 736b8e5a13f385c1bbac3359e8a86c3f58632751d9b57af133cd6434b5c9e33a Dec 10 19:07:51 crc kubenswrapper[4828]: I1210 19:07:51.230869 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:07:51 crc kubenswrapper[4828]: I1210 19:07:51.230923 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:07:52 crc kubenswrapper[4828]: I1210 19:07:52.026384 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-57l2h" event={"ID":"ebb4036a-1bd8-4636-8bea-63b0a6b4cc38","Type":"ContainerStarted","Data":"736b8e5a13f385c1bbac3359e8a86c3f58632751d9b57af133cd6434b5c9e33a"} Dec 10 19:07:53 crc kubenswrapper[4828]: I1210 19:07:53.818467 4828 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","pod1f37f60f-fe12-4530-8ff1-ea3794c65597"] err="unable to destroy cgroup paths for cgroup [kubepods burstable pod1f37f60f-fe12-4530-8ff1-ea3794c65597] : Timed out while waiting for systemd to remove kubepods-burstable-pod1f37f60f_fe12_4530_8ff1_ea3794c65597.slice" Dec 10 19:07:54 crc kubenswrapper[4828]: I1210 19:07:54.036913 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-vwlbb" Dec 10 19:08:03 crc kubenswrapper[4828]: E1210 19:08:03.344241 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec" Dec 10 19:08:03 crc kubenswrapper[4828]: E1210 19:08:03.344994 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-6774fff8-2glrf_openshift-operators(68b3a59c-1ec9-4381-a92a-810b554d76e0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 19:08:03 crc kubenswrapper[4828]: E1210 19:08:03.346154 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" podUID="68b3a59c-1ec9-4381-a92a-810b554d76e0" Dec 10 19:08:04 crc kubenswrapper[4828]: E1210 19:08:04.105344 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:43d33f0125e6b990f4a972ac4e952a065d7e72dc1690c6c836963b7341734aec\\\"\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" podUID="68b3a59c-1ec9-4381-a92a-810b554d76e0" Dec 10 19:08:04 crc kubenswrapper[4828]: E1210 19:08:04.973852 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb" Dec 10 19:08:04 crc kubenswrapper[4828]: E1210 19:08:04.974218 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb,Command:[],Args:[--namespace=$(NAMESPACE) --images=perses=$(RELATED_IMAGE_PERSES) --images=alertmanager=$(RELATED_IMAGE_ALERTMANAGER) --images=prometheus=$(RELATED_IMAGE_PROMETHEUS) --images=thanos=$(RELATED_IMAGE_THANOS) --images=ui-dashboards=$(RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN) --images=ui-distributed-tracing=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN) --images=ui-distributed-tracing-pf5=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5) --images=ui-distributed-tracing-pf4=$(RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4) --images=ui-logging=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN) --images=ui-logging-pf4=$(RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4) --images=ui-troubleshooting-panel=$(RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN) --images=ui-monitoring=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN) --images=ui-monitoring-pf5=$(RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5) --images=korrel8r=$(RELATED_IMAGE_KORREL8R) --images=health-analyzer=$(RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER) --openshift.enabled=true],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:RELATED_IMAGE_ALERTMANAGER,Value:registry.redhat.io/cluster-observability-operator/alertmanager-rhel9@sha256:e718854a7d6ca8accf0fa72db0eb902e46c44d747ad51dc3f06bba0cefaa3c01,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS,Value:registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:17ea20be390a94ab39f5cdd7f0cbc2498046eebcf77fe3dec9aa288d5c2cf46b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_THANOS,Value:registry.redhat.io/cluster-observability-operator/thanos-rhel9@sha256:d972f4faa5e9c121402d23ed85002f26af48ec36b1b71a7489d677b3913d08b4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PERSES,Value:registry.redhat.io/cluster-observability-operator/perses-rhel9@sha256:91531137fc1dcd740e277e0f65e120a0176a16f788c14c27925b61aa0b792ade,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DASHBOARDS_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/dashboards-console-plugin-rhel9@sha256:a69da8bbca8a28dd2925f864d51cc31cf761b10532c553095ba40b242ef701cb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-rhel9@sha256:897e1bfad1187062725b54d87107bd0155972257a50d8335dd29e1999b828a4f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-pf5-rhel9@sha256:95fe5b5746ca8c07ac9217ce2d8ac8e6afad17af210f9d8e0074df1310b209a8,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_DISTRIBUTED_TRACING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/distributed-tracing-console-plugin-pf4-rhel9@sha256:e9d9a89e4d8126a62b1852055482258ee528cac6398dd5d43ebad75ace0f33c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-rhel9@sha256:ec684a0645ceb917b019af7ddba68c3533416e356ab0d0320a30e75ca7ebb31b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_LOGGING_PLUGIN_PF4,Value:registry.redhat.io/cluster-observability-operator/logging-console-plugin-pf4-rhel9@sha256:3b9693fcde9b3a9494fb04735b1f7cfd0426f10be820fdc3f024175c0d3df1c9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_TROUBLESHOOTING_PANEL_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/troubleshooting-panel-console-plugin-rhel9@sha256:580606f194180accc8abba099e17a26dca7522ec6d233fa2fdd40312771703e3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-rhel9@sha256:e03777be39e71701935059cd877603874a13ac94daa73219d4e5e545599d78a9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CONSOLE_MONITORING_PLUGIN_PF5,Value:registry.redhat.io/cluster-observability-operator/monitoring-console-plugin-pf5-rhel9@sha256:aa47256193cfd2877853878e1ae97d2ab8b8e5deae62b387cbfad02b284d379c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KORREL8R,Value:registry.redhat.io/cluster-observability-operator/korrel8r-rhel9@sha256:c595ff56b2cb85514bf4784db6ddb82e4e657e3e708a7fb695fc4997379a94d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLUSTER_HEALTH_ANALYZER,Value:registry.redhat.io/cluster-observability-operator/cluster-health-analyzer-rhel9@sha256:45a4ec2a519bcec99e886aa91596d5356a2414a2bd103baaef9fa7838c672eb2,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{400 -3} {} 400m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:observability-operator-tls,ReadOnly:true,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dmgtc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod observability-operator-d8bb48f5d-flqjw_openshift-operators(b55b0a41-ae3f-4acf-bf2d-0378868dca03): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 10 19:08:04 crc kubenswrapper[4828]: E1210 19:08:04.975432 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" podUID="b55b0a41-ae3f-4acf-bf2d-0378868dca03" Dec 10 19:08:05 crc kubenswrapper[4828]: E1210 19:08:05.109993 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/cluster-observability-rhel9-operator@sha256:ce7d2904f7b238aa37dfe74a0b76bf73629e7a14fa52bf54b0ecf030ca36f1bb\\\"\"" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" podUID="b55b0a41-ae3f-4acf-bf2d-0378868dca03" Dec 10 19:08:06 crc kubenswrapper[4828]: I1210 19:08:06.114058 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" event={"ID":"b5194787-06b3-40a3-92dd-7a7471ba5829","Type":"ContainerStarted","Data":"3aca4d792b457837af933f4beadd82a29a0ed5e661fad01ff78651877b038912"} Dec 10 19:08:06 crc kubenswrapper[4828]: I1210 19:08:06.115706 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" event={"ID":"fe95ea54-da09-4ec9-a930-5be068b09954","Type":"ContainerStarted","Data":"0adb8a11f9904ea1045ee3797b995780015ff0a1d84a6dde4214f31cf88275ea"} Dec 10 19:08:06 crc kubenswrapper[4828]: I1210 19:08:06.117288 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-57l2h" event={"ID":"ebb4036a-1bd8-4636-8bea-63b0a6b4cc38","Type":"ContainerStarted","Data":"962632ee6a5b60586991b7253b66be0afe6e8bf4588dc7b20f182474ac7aecf0"} Dec 10 19:08:06 crc kubenswrapper[4828]: I1210 19:08:06.117468 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:08:06 crc kubenswrapper[4828]: I1210 19:08:06.134617 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-sbnvz" podStartSLOduration=16.466053778 podStartE2EDuration="33.134592375s" podCreationTimestamp="2025-12-10 19:07:33 +0000 UTC" firstStartedPulling="2025-12-10 19:07:48.28545661 +0000 UTC m=+748.796067615" lastFinishedPulling="2025-12-10 19:08:04.953995207 +0000 UTC m=+765.464606212" observedRunningTime="2025-12-10 19:08:06.131045434 +0000 UTC m=+766.641656449" watchObservedRunningTime="2025-12-10 19:08:06.134592375 +0000 UTC m=+766.645203390" Dec 10 19:08:06 crc kubenswrapper[4828]: I1210 19:08:06.160102 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-vpjkk" podStartSLOduration=18.417536016 podStartE2EDuration="33.160078096s" podCreationTimestamp="2025-12-10 19:07:33 +0000 UTC" firstStartedPulling="2025-12-10 19:07:50.226001889 +0000 UTC m=+750.736612894" lastFinishedPulling="2025-12-10 19:08:04.968543969 +0000 UTC m=+765.479154974" observedRunningTime="2025-12-10 19:08:06.153071936 +0000 UTC m=+766.663682971" watchObservedRunningTime="2025-12-10 19:08:06.160078096 +0000 UTC m=+766.670689101" Dec 10 19:08:06 crc kubenswrapper[4828]: I1210 19:08:06.181073 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-57l2h" podStartSLOduration=19.40946523 podStartE2EDuration="33.181050015s" podCreationTimestamp="2025-12-10 19:07:33 +0000 UTC" firstStartedPulling="2025-12-10 19:07:51.199725357 +0000 UTC m=+751.710336352" lastFinishedPulling="2025-12-10 19:08:04.971310132 +0000 UTC m=+765.481921137" observedRunningTime="2025-12-10 19:08:06.1769128 +0000 UTC m=+766.687523815" watchObservedRunningTime="2025-12-10 19:08:06.181050015 +0000 UTC m=+766.691661020" Dec 10 19:08:14 crc kubenswrapper[4828]: I1210 19:08:14.076334 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-57l2h" Dec 10 19:08:19 crc kubenswrapper[4828]: I1210 19:08:19.204970 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" event={"ID":"68b3a59c-1ec9-4381-a92a-810b554d76e0","Type":"ContainerStarted","Data":"ae96dd6cfc3b369de0f008df4c3f1036472b8c5802034d83aa89ef2246084ee1"} Dec 10 19:08:19 crc kubenswrapper[4828]: I1210 19:08:19.231966 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-6774fff8-2glrf" podStartSLOduration=-9223371990.622839 podStartE2EDuration="46.231937067s" podCreationTimestamp="2025-12-10 19:07:33 +0000 UTC" firstStartedPulling="2025-12-10 19:07:47.107920462 +0000 UTC m=+747.618531477" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:08:19.220468975 +0000 UTC m=+779.731080000" watchObservedRunningTime="2025-12-10 19:08:19.231937067 +0000 UTC m=+779.742548072" Dec 10 19:08:20 crc kubenswrapper[4828]: I1210 19:08:20.213094 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" event={"ID":"b55b0a41-ae3f-4acf-bf2d-0378868dca03","Type":"ContainerStarted","Data":"dcfc37c80d395382b3c64ec1b59ab3295048049b7ee50d5de21035871e8e2966"} Dec 10 19:08:20 crc kubenswrapper[4828]: I1210 19:08:20.213952 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:08:20 crc kubenswrapper[4828]: I1210 19:08:20.216016 4828 patch_prober.go:28] interesting pod/observability-operator-d8bb48f5d-flqjw container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.16:8081/healthz\": dial tcp 10.217.0.16:8081: connect: connection refused" start-of-body= Dec 10 19:08:20 crc kubenswrapper[4828]: I1210 19:08:20.216060 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" podUID="b55b0a41-ae3f-4acf-bf2d-0378868dca03" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.16:8081/healthz\": dial tcp 10.217.0.16:8081: connect: connection refused" Dec 10 19:08:20 crc kubenswrapper[4828]: I1210 19:08:20.235606 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" podStartSLOduration=17.080967135 podStartE2EDuration="47.235573957s" podCreationTimestamp="2025-12-10 19:07:33 +0000 UTC" firstStartedPulling="2025-12-10 19:07:49.737911622 +0000 UTC m=+750.248522637" lastFinishedPulling="2025-12-10 19:08:19.892518444 +0000 UTC m=+780.403129459" observedRunningTime="2025-12-10 19:08:20.234602862 +0000 UTC m=+780.745213867" watchObservedRunningTime="2025-12-10 19:08:20.235573957 +0000 UTC m=+780.746184962" Dec 10 19:08:21 crc kubenswrapper[4828]: I1210 19:08:21.219364 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-flqjw" Dec 10 19:08:21 crc kubenswrapper[4828]: I1210 19:08:21.230425 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:08:21 crc kubenswrapper[4828]: I1210 19:08:21.230488 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:08:21 crc kubenswrapper[4828]: I1210 19:08:21.230529 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 19:08:21 crc kubenswrapper[4828]: I1210 19:08:21.231088 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1d400e5fe2702c739b92ea52b2e9453f005b08c30f146bdcec9d4d95ba2cb93b"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:08:21 crc kubenswrapper[4828]: I1210 19:08:21.231148 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://1d400e5fe2702c739b92ea52b2e9453f005b08c30f146bdcec9d4d95ba2cb93b" gracePeriod=600 Dec 10 19:08:22 crc kubenswrapper[4828]: I1210 19:08:22.225836 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="1d400e5fe2702c739b92ea52b2e9453f005b08c30f146bdcec9d4d95ba2cb93b" exitCode=0 Dec 10 19:08:22 crc kubenswrapper[4828]: I1210 19:08:22.225912 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"1d400e5fe2702c739b92ea52b2e9453f005b08c30f146bdcec9d4d95ba2cb93b"} Dec 10 19:08:22 crc kubenswrapper[4828]: I1210 19:08:22.226348 4828 scope.go:117] "RemoveContainer" containerID="9e2e9020bcb1e1466a5a79b6ec9f630360e0d3e7e570a3abeb544c0a74e30c14" Dec 10 19:08:23 crc kubenswrapper[4828]: I1210 19:08:23.233849 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"97c29c3bfdeed7777dc023da2b04e9fc93587d77035f3f054b943162232fe57b"} Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.113890 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-x7cdj"] Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.116001 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-x7cdj" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.119975 4828 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-5lf4j" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.120250 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.122227 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.131389 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-x7cdj"] Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.144698 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nlz6n"] Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.145463 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-nlz6n" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.161624 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nlz6n"] Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.163639 4828 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-mx8jq" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.171016 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kmxf\" (UniqueName: \"kubernetes.io/projected/eaad2b2e-c77d-4f5b-a4e7-b6fe4ae46e3a-kube-api-access-5kmxf\") pod \"cert-manager-cainjector-7f985d654d-x7cdj\" (UID: \"eaad2b2e-c77d-4f5b-a4e7-b6fe4ae46e3a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-x7cdj" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.171133 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw9tt\" (UniqueName: \"kubernetes.io/projected/45d36af5-9592-4057-a3fa-00c9cd1aa9e5-kube-api-access-dw9tt\") pod \"cert-manager-5b446d88c5-nlz6n\" (UID: \"45d36af5-9592-4057-a3fa-00c9cd1aa9e5\") " pod="cert-manager/cert-manager-5b446d88c5-nlz6n" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.173123 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-gtg7k"] Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.173987 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-gtg7k" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.183436 4828 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-qcfrc" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.196355 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-gtg7k"] Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.272146 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw9tt\" (UniqueName: \"kubernetes.io/projected/45d36af5-9592-4057-a3fa-00c9cd1aa9e5-kube-api-access-dw9tt\") pod \"cert-manager-5b446d88c5-nlz6n\" (UID: \"45d36af5-9592-4057-a3fa-00c9cd1aa9e5\") " pod="cert-manager/cert-manager-5b446d88c5-nlz6n" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.272210 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kmxf\" (UniqueName: \"kubernetes.io/projected/eaad2b2e-c77d-4f5b-a4e7-b6fe4ae46e3a-kube-api-access-5kmxf\") pod \"cert-manager-cainjector-7f985d654d-x7cdj\" (UID: \"eaad2b2e-c77d-4f5b-a4e7-b6fe4ae46e3a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-x7cdj" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.290899 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw9tt\" (UniqueName: \"kubernetes.io/projected/45d36af5-9592-4057-a3fa-00c9cd1aa9e5-kube-api-access-dw9tt\") pod \"cert-manager-5b446d88c5-nlz6n\" (UID: \"45d36af5-9592-4057-a3fa-00c9cd1aa9e5\") " pod="cert-manager/cert-manager-5b446d88c5-nlz6n" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.290983 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kmxf\" (UniqueName: \"kubernetes.io/projected/eaad2b2e-c77d-4f5b-a4e7-b6fe4ae46e3a-kube-api-access-5kmxf\") pod \"cert-manager-cainjector-7f985d654d-x7cdj\" (UID: \"eaad2b2e-c77d-4f5b-a4e7-b6fe4ae46e3a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-x7cdj" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.374217 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ffpz\" (UniqueName: \"kubernetes.io/projected/6e73ff05-cad9-439a-9116-c885059594f2-kube-api-access-8ffpz\") pod \"cert-manager-webhook-5655c58dd6-gtg7k\" (UID: \"6e73ff05-cad9-439a-9116-c885059594f2\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-gtg7k" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.431889 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-x7cdj" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.463272 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-nlz6n" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.475093 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ffpz\" (UniqueName: \"kubernetes.io/projected/6e73ff05-cad9-439a-9116-c885059594f2-kube-api-access-8ffpz\") pod \"cert-manager-webhook-5655c58dd6-gtg7k\" (UID: \"6e73ff05-cad9-439a-9116-c885059594f2\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-gtg7k" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.500581 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ffpz\" (UniqueName: \"kubernetes.io/projected/6e73ff05-cad9-439a-9116-c885059594f2-kube-api-access-8ffpz\") pod \"cert-manager-webhook-5655c58dd6-gtg7k\" (UID: \"6e73ff05-cad9-439a-9116-c885059594f2\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-gtg7k" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.631186 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-x7cdj"] Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.688401 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-nlz6n"] Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.789241 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-gtg7k" Dec 10 19:08:29 crc kubenswrapper[4828]: I1210 19:08:29.988949 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-gtg7k"] Dec 10 19:08:29 crc kubenswrapper[4828]: W1210 19:08:29.992540 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e73ff05_cad9_439a_9116_c885059594f2.slice/crio-958d3e7e629c3cc862a4f690bc71c9a97fd61378d379ef8014f8769a9f88e74a WatchSource:0}: Error finding container 958d3e7e629c3cc862a4f690bc71c9a97fd61378d379ef8014f8769a9f88e74a: Status 404 returned error can't find the container with id 958d3e7e629c3cc862a4f690bc71c9a97fd61378d379ef8014f8769a9f88e74a Dec 10 19:08:30 crc kubenswrapper[4828]: I1210 19:08:30.276341 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-nlz6n" event={"ID":"45d36af5-9592-4057-a3fa-00c9cd1aa9e5","Type":"ContainerStarted","Data":"b162b4371b642240f1c0c70d40e8e4dd8df50dc415dfa0d82680cda731fcb5f7"} Dec 10 19:08:30 crc kubenswrapper[4828]: I1210 19:08:30.278410 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-gtg7k" event={"ID":"6e73ff05-cad9-439a-9116-c885059594f2","Type":"ContainerStarted","Data":"958d3e7e629c3cc862a4f690bc71c9a97fd61378d379ef8014f8769a9f88e74a"} Dec 10 19:08:30 crc kubenswrapper[4828]: I1210 19:08:30.279772 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-x7cdj" event={"ID":"eaad2b2e-c77d-4f5b-a4e7-b6fe4ae46e3a","Type":"ContainerStarted","Data":"5e97479b64705d0e68f7a77fe96024793cda85f76726be86e03f745101ebb0b3"} Dec 10 19:08:35 crc kubenswrapper[4828]: I1210 19:08:35.324473 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-x7cdj" event={"ID":"eaad2b2e-c77d-4f5b-a4e7-b6fe4ae46e3a","Type":"ContainerStarted","Data":"13fac9cee09675ed965bc2cce4cc431301d294181d13feb87d73beb88fd1a210"} Dec 10 19:08:35 crc kubenswrapper[4828]: I1210 19:08:35.327442 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-nlz6n" event={"ID":"45d36af5-9592-4057-a3fa-00c9cd1aa9e5","Type":"ContainerStarted","Data":"c359c6b4368a054a206a0959d3ee95654572f230bf2481378c0958518ce319a1"} Dec 10 19:08:35 crc kubenswrapper[4828]: I1210 19:08:35.342537 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-x7cdj" podStartSLOduration=1.4356058809999999 podStartE2EDuration="6.342521923s" podCreationTimestamp="2025-12-10 19:08:29 +0000 UTC" firstStartedPulling="2025-12-10 19:08:29.64320844 +0000 UTC m=+790.153819445" lastFinishedPulling="2025-12-10 19:08:34.550124482 +0000 UTC m=+795.060735487" observedRunningTime="2025-12-10 19:08:35.33930067 +0000 UTC m=+795.849911695" watchObservedRunningTime="2025-12-10 19:08:35.342521923 +0000 UTC m=+795.853132928" Dec 10 19:08:35 crc kubenswrapper[4828]: I1210 19:08:35.353430 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-nlz6n" podStartSLOduration=1.444161828 podStartE2EDuration="6.35340979s" podCreationTimestamp="2025-12-10 19:08:29 +0000 UTC" firstStartedPulling="2025-12-10 19:08:29.69142343 +0000 UTC m=+790.202034435" lastFinishedPulling="2025-12-10 19:08:34.600671382 +0000 UTC m=+795.111282397" observedRunningTime="2025-12-10 19:08:35.350740652 +0000 UTC m=+795.861351657" watchObservedRunningTime="2025-12-10 19:08:35.35340979 +0000 UTC m=+795.864020805" Dec 10 19:08:36 crc kubenswrapper[4828]: I1210 19:08:36.334749 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-gtg7k" event={"ID":"6e73ff05-cad9-439a-9116-c885059594f2","Type":"ContainerStarted","Data":"2b0fe9517181026fc996eddee6f2566255154a64867c36d0bee003b710f839bb"} Dec 10 19:08:36 crc kubenswrapper[4828]: I1210 19:08:36.335409 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-gtg7k" Dec 10 19:08:36 crc kubenswrapper[4828]: I1210 19:08:36.349266 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-gtg7k" podStartSLOduration=1.677853152 podStartE2EDuration="7.349247062s" podCreationTimestamp="2025-12-10 19:08:29 +0000 UTC" firstStartedPulling="2025-12-10 19:08:29.99512372 +0000 UTC m=+790.505734725" lastFinishedPulling="2025-12-10 19:08:35.66651763 +0000 UTC m=+796.177128635" observedRunningTime="2025-12-10 19:08:36.348738609 +0000 UTC m=+796.859349614" watchObservedRunningTime="2025-12-10 19:08:36.349247062 +0000 UTC m=+796.859858077" Dec 10 19:08:44 crc kubenswrapper[4828]: I1210 19:08:44.792885 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-gtg7k" Dec 10 19:09:11 crc kubenswrapper[4828]: I1210 19:09:11.904171 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5"] Dec 10 19:09:11 crc kubenswrapper[4828]: I1210 19:09:11.905946 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" Dec 10 19:09:11 crc kubenswrapper[4828]: I1210 19:09:11.917229 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5"] Dec 10 19:09:11 crc kubenswrapper[4828]: I1210 19:09:11.917915 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.000251 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef8a4983-2aae-4e7e-a416-317ffceb7b53-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5\" (UID: \"ef8a4983-2aae-4e7e-a416-317ffceb7b53\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.000558 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jf65\" (UniqueName: \"kubernetes.io/projected/ef8a4983-2aae-4e7e-a416-317ffceb7b53-kube-api-access-2jf65\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5\" (UID: \"ef8a4983-2aae-4e7e-a416-317ffceb7b53\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.000929 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef8a4983-2aae-4e7e-a416-317ffceb7b53-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5\" (UID: \"ef8a4983-2aae-4e7e-a416-317ffceb7b53\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.101929 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef8a4983-2aae-4e7e-a416-317ffceb7b53-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5\" (UID: \"ef8a4983-2aae-4e7e-a416-317ffceb7b53\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.101984 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef8a4983-2aae-4e7e-a416-317ffceb7b53-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5\" (UID: \"ef8a4983-2aae-4e7e-a416-317ffceb7b53\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.102003 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jf65\" (UniqueName: \"kubernetes.io/projected/ef8a4983-2aae-4e7e-a416-317ffceb7b53-kube-api-access-2jf65\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5\" (UID: \"ef8a4983-2aae-4e7e-a416-317ffceb7b53\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.102648 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef8a4983-2aae-4e7e-a416-317ffceb7b53-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5\" (UID: \"ef8a4983-2aae-4e7e-a416-317ffceb7b53\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.102732 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef8a4983-2aae-4e7e-a416-317ffceb7b53-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5\" (UID: \"ef8a4983-2aae-4e7e-a416-317ffceb7b53\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.110890 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws"] Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.112548 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.123368 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws"] Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.127455 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jf65\" (UniqueName: \"kubernetes.io/projected/ef8a4983-2aae-4e7e-a416-317ffceb7b53-kube-api-access-2jf65\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5\" (UID: \"ef8a4983-2aae-4e7e-a416-317ffceb7b53\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.203510 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws\" (UID: \"2c2bf8e7-87cf-4ede-bc49-e955290a66ee\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.203743 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvc8q\" (UniqueName: \"kubernetes.io/projected/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-kube-api-access-qvc8q\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws\" (UID: \"2c2bf8e7-87cf-4ede-bc49-e955290a66ee\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.203897 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws\" (UID: \"2c2bf8e7-87cf-4ede-bc49-e955290a66ee\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.226310 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.304756 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvc8q\" (UniqueName: \"kubernetes.io/projected/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-kube-api-access-qvc8q\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws\" (UID: \"2c2bf8e7-87cf-4ede-bc49-e955290a66ee\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.305050 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws\" (UID: \"2c2bf8e7-87cf-4ede-bc49-e955290a66ee\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.305079 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws\" (UID: \"2c2bf8e7-87cf-4ede-bc49-e955290a66ee\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.305530 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws\" (UID: \"2c2bf8e7-87cf-4ede-bc49-e955290a66ee\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.305756 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws\" (UID: \"2c2bf8e7-87cf-4ede-bc49-e955290a66ee\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.327123 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvc8q\" (UniqueName: \"kubernetes.io/projected/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-kube-api-access-qvc8q\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws\" (UID: \"2c2bf8e7-87cf-4ede-bc49-e955290a66ee\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.450225 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.625705 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5"] Dec 10 19:09:12 crc kubenswrapper[4828]: W1210 19:09:12.630132 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef8a4983_2aae_4e7e_a416_317ffceb7b53.slice/crio-89fe9acdaee267439be2d9fa74dfa7e96d6af84472f2802d41fa85fe0c1c8cf2 WatchSource:0}: Error finding container 89fe9acdaee267439be2d9fa74dfa7e96d6af84472f2802d41fa85fe0c1c8cf2: Status 404 returned error can't find the container with id 89fe9acdaee267439be2d9fa74dfa7e96d6af84472f2802d41fa85fe0c1c8cf2 Dec 10 19:09:12 crc kubenswrapper[4828]: I1210 19:09:12.844121 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws"] Dec 10 19:09:13 crc kubenswrapper[4828]: I1210 19:09:13.572823 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" event={"ID":"2c2bf8e7-87cf-4ede-bc49-e955290a66ee","Type":"ContainerStarted","Data":"0045f5fc8870cca613d1a05398debe2d46c8b1fca11fc69dcc153fafeed0a67b"} Dec 10 19:09:13 crc kubenswrapper[4828]: I1210 19:09:13.575710 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" event={"ID":"ef8a4983-2aae-4e7e-a416-317ffceb7b53","Type":"ContainerStarted","Data":"89fe9acdaee267439be2d9fa74dfa7e96d6af84472f2802d41fa85fe0c1c8cf2"} Dec 10 19:09:14 crc kubenswrapper[4828]: I1210 19:09:14.586660 4828 generic.go:334] "Generic (PLEG): container finished" podID="2c2bf8e7-87cf-4ede-bc49-e955290a66ee" containerID="4d984e11ed5cf91b19dfef5a50f07097c6dd4d9089721fcd234b7031c9a3d300" exitCode=0 Dec 10 19:09:14 crc kubenswrapper[4828]: I1210 19:09:14.586733 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" event={"ID":"2c2bf8e7-87cf-4ede-bc49-e955290a66ee","Type":"ContainerDied","Data":"4d984e11ed5cf91b19dfef5a50f07097c6dd4d9089721fcd234b7031c9a3d300"} Dec 10 19:09:14 crc kubenswrapper[4828]: I1210 19:09:14.588571 4828 generic.go:334] "Generic (PLEG): container finished" podID="ef8a4983-2aae-4e7e-a416-317ffceb7b53" containerID="5a5b2354cbe76d2c21357d9068fdd80762ef62956fc677e4f7b9eddf04c06aab" exitCode=0 Dec 10 19:09:14 crc kubenswrapper[4828]: I1210 19:09:14.588638 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" event={"ID":"ef8a4983-2aae-4e7e-a416-317ffceb7b53","Type":"ContainerDied","Data":"5a5b2354cbe76d2c21357d9068fdd80762ef62956fc677e4f7b9eddf04c06aab"} Dec 10 19:09:15 crc kubenswrapper[4828]: I1210 19:09:15.663769 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-grpgt"] Dec 10 19:09:15 crc kubenswrapper[4828]: I1210 19:09:15.665480 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-grpgt" Dec 10 19:09:15 crc kubenswrapper[4828]: I1210 19:09:15.676898 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-grpgt"] Dec 10 19:09:15 crc kubenswrapper[4828]: I1210 19:09:15.753282 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/288af64f-b8e1-4597-ab58-8fb61bf0d13c-utilities\") pod \"redhat-operators-grpgt\" (UID: \"288af64f-b8e1-4597-ab58-8fb61bf0d13c\") " pod="openshift-marketplace/redhat-operators-grpgt" Dec 10 19:09:15 crc kubenswrapper[4828]: I1210 19:09:15.753493 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/288af64f-b8e1-4597-ab58-8fb61bf0d13c-catalog-content\") pod \"redhat-operators-grpgt\" (UID: \"288af64f-b8e1-4597-ab58-8fb61bf0d13c\") " pod="openshift-marketplace/redhat-operators-grpgt" Dec 10 19:09:15 crc kubenswrapper[4828]: I1210 19:09:15.753760 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrb7k\" (UniqueName: \"kubernetes.io/projected/288af64f-b8e1-4597-ab58-8fb61bf0d13c-kube-api-access-jrb7k\") pod \"redhat-operators-grpgt\" (UID: \"288af64f-b8e1-4597-ab58-8fb61bf0d13c\") " pod="openshift-marketplace/redhat-operators-grpgt" Dec 10 19:09:15 crc kubenswrapper[4828]: I1210 19:09:15.855464 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrb7k\" (UniqueName: \"kubernetes.io/projected/288af64f-b8e1-4597-ab58-8fb61bf0d13c-kube-api-access-jrb7k\") pod \"redhat-operators-grpgt\" (UID: \"288af64f-b8e1-4597-ab58-8fb61bf0d13c\") " pod="openshift-marketplace/redhat-operators-grpgt" Dec 10 19:09:15 crc kubenswrapper[4828]: I1210 19:09:15.855519 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/288af64f-b8e1-4597-ab58-8fb61bf0d13c-utilities\") pod \"redhat-operators-grpgt\" (UID: \"288af64f-b8e1-4597-ab58-8fb61bf0d13c\") " pod="openshift-marketplace/redhat-operators-grpgt" Dec 10 19:09:15 crc kubenswrapper[4828]: I1210 19:09:15.855574 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/288af64f-b8e1-4597-ab58-8fb61bf0d13c-catalog-content\") pod \"redhat-operators-grpgt\" (UID: \"288af64f-b8e1-4597-ab58-8fb61bf0d13c\") " pod="openshift-marketplace/redhat-operators-grpgt" Dec 10 19:09:15 crc kubenswrapper[4828]: I1210 19:09:15.856023 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/288af64f-b8e1-4597-ab58-8fb61bf0d13c-utilities\") pod \"redhat-operators-grpgt\" (UID: \"288af64f-b8e1-4597-ab58-8fb61bf0d13c\") " pod="openshift-marketplace/redhat-operators-grpgt" Dec 10 19:09:15 crc kubenswrapper[4828]: I1210 19:09:15.856070 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/288af64f-b8e1-4597-ab58-8fb61bf0d13c-catalog-content\") pod \"redhat-operators-grpgt\" (UID: \"288af64f-b8e1-4597-ab58-8fb61bf0d13c\") " pod="openshift-marketplace/redhat-operators-grpgt" Dec 10 19:09:15 crc kubenswrapper[4828]: I1210 19:09:15.875346 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrb7k\" (UniqueName: \"kubernetes.io/projected/288af64f-b8e1-4597-ab58-8fb61bf0d13c-kube-api-access-jrb7k\") pod \"redhat-operators-grpgt\" (UID: \"288af64f-b8e1-4597-ab58-8fb61bf0d13c\") " pod="openshift-marketplace/redhat-operators-grpgt" Dec 10 19:09:15 crc kubenswrapper[4828]: I1210 19:09:15.991586 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-grpgt" Dec 10 19:09:18 crc kubenswrapper[4828]: I1210 19:09:18.283116 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-grpgt"] Dec 10 19:09:18 crc kubenswrapper[4828]: W1210 19:09:18.288535 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod288af64f_b8e1_4597_ab58_8fb61bf0d13c.slice/crio-f75547a606f98737a864ab17e97987385e6773f62b8b352165a0144864f170b0 WatchSource:0}: Error finding container f75547a606f98737a864ab17e97987385e6773f62b8b352165a0144864f170b0: Status 404 returned error can't find the container with id f75547a606f98737a864ab17e97987385e6773f62b8b352165a0144864f170b0 Dec 10 19:09:18 crc kubenswrapper[4828]: I1210 19:09:18.614227 4828 generic.go:334] "Generic (PLEG): container finished" podID="ef8a4983-2aae-4e7e-a416-317ffceb7b53" containerID="0371a30101d382bb9eef91b4ddb622d651413800d93d23c796121b8b07709122" exitCode=0 Dec 10 19:09:18 crc kubenswrapper[4828]: I1210 19:09:18.614336 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" event={"ID":"ef8a4983-2aae-4e7e-a416-317ffceb7b53","Type":"ContainerDied","Data":"0371a30101d382bb9eef91b4ddb622d651413800d93d23c796121b8b07709122"} Dec 10 19:09:18 crc kubenswrapper[4828]: I1210 19:09:18.616638 4828 generic.go:334] "Generic (PLEG): container finished" podID="288af64f-b8e1-4597-ab58-8fb61bf0d13c" containerID="61b0dd212618f1bbf2d7229a16df935f3b48abce99b3f5504c6d47cd34e10608" exitCode=0 Dec 10 19:09:18 crc kubenswrapper[4828]: I1210 19:09:18.616708 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grpgt" event={"ID":"288af64f-b8e1-4597-ab58-8fb61bf0d13c","Type":"ContainerDied","Data":"61b0dd212618f1bbf2d7229a16df935f3b48abce99b3f5504c6d47cd34e10608"} Dec 10 19:09:18 crc kubenswrapper[4828]: I1210 19:09:18.616757 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grpgt" event={"ID":"288af64f-b8e1-4597-ab58-8fb61bf0d13c","Type":"ContainerStarted","Data":"f75547a606f98737a864ab17e97987385e6773f62b8b352165a0144864f170b0"} Dec 10 19:09:18 crc kubenswrapper[4828]: I1210 19:09:18.621902 4828 generic.go:334] "Generic (PLEG): container finished" podID="2c2bf8e7-87cf-4ede-bc49-e955290a66ee" containerID="2b7ae35d07f2ead19b61ba3e7857b464914c62d1c9beb6178cf3f4eb3c5fe008" exitCode=0 Dec 10 19:09:18 crc kubenswrapper[4828]: I1210 19:09:18.621961 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" event={"ID":"2c2bf8e7-87cf-4ede-bc49-e955290a66ee","Type":"ContainerDied","Data":"2b7ae35d07f2ead19b61ba3e7857b464914c62d1c9beb6178cf3f4eb3c5fe008"} Dec 10 19:09:19 crc kubenswrapper[4828]: I1210 19:09:19.630608 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" event={"ID":"2c2bf8e7-87cf-4ede-bc49-e955290a66ee","Type":"ContainerStarted","Data":"c5ee05dc0867f99d12548519a59a9a4b0ea8e058347245c6bf6d4d5e0a53c1d9"} Dec 10 19:09:19 crc kubenswrapper[4828]: I1210 19:09:19.632791 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" event={"ID":"ef8a4983-2aae-4e7e-a416-317ffceb7b53","Type":"ContainerStarted","Data":"b19f1b5034590d1301f78db875436ea800cca53cc1b7466da8d58f24cfbaa75a"} Dec 10 19:09:19 crc kubenswrapper[4828]: I1210 19:09:19.655212 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" podStartSLOduration=4.378372045 podStartE2EDuration="7.655194572s" podCreationTimestamp="2025-12-10 19:09:12 +0000 UTC" firstStartedPulling="2025-12-10 19:09:14.589049965 +0000 UTC m=+835.099660970" lastFinishedPulling="2025-12-10 19:09:17.865872502 +0000 UTC m=+838.376483497" observedRunningTime="2025-12-10 19:09:19.649777374 +0000 UTC m=+840.160388379" watchObservedRunningTime="2025-12-10 19:09:19.655194572 +0000 UTC m=+840.165805587" Dec 10 19:09:19 crc kubenswrapper[4828]: I1210 19:09:19.671636 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" podStartSLOduration=5.396750954 podStartE2EDuration="8.671617381s" podCreationTimestamp="2025-12-10 19:09:11 +0000 UTC" firstStartedPulling="2025-12-10 19:09:14.591076057 +0000 UTC m=+835.101687102" lastFinishedPulling="2025-12-10 19:09:17.865942524 +0000 UTC m=+838.376553529" observedRunningTime="2025-12-10 19:09:19.666929472 +0000 UTC m=+840.177540487" watchObservedRunningTime="2025-12-10 19:09:19.671617381 +0000 UTC m=+840.182228386" Dec 10 19:09:20 crc kubenswrapper[4828]: I1210 19:09:20.643975 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grpgt" event={"ID":"288af64f-b8e1-4597-ab58-8fb61bf0d13c","Type":"ContainerStarted","Data":"3e54a0b3b54fd443dfeccf860b359b18ebe7d0bde4b93ac27f3b5e4ca13d5bb0"} Dec 10 19:09:20 crc kubenswrapper[4828]: I1210 19:09:20.647544 4828 generic.go:334] "Generic (PLEG): container finished" podID="2c2bf8e7-87cf-4ede-bc49-e955290a66ee" containerID="c5ee05dc0867f99d12548519a59a9a4b0ea8e058347245c6bf6d4d5e0a53c1d9" exitCode=0 Dec 10 19:09:20 crc kubenswrapper[4828]: I1210 19:09:20.647657 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" event={"ID":"2c2bf8e7-87cf-4ede-bc49-e955290a66ee","Type":"ContainerDied","Data":"c5ee05dc0867f99d12548519a59a9a4b0ea8e058347245c6bf6d4d5e0a53c1d9"} Dec 10 19:09:20 crc kubenswrapper[4828]: I1210 19:09:20.651699 4828 generic.go:334] "Generic (PLEG): container finished" podID="ef8a4983-2aae-4e7e-a416-317ffceb7b53" containerID="b19f1b5034590d1301f78db875436ea800cca53cc1b7466da8d58f24cfbaa75a" exitCode=0 Dec 10 19:09:20 crc kubenswrapper[4828]: I1210 19:09:20.651769 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" event={"ID":"ef8a4983-2aae-4e7e-a416-317ffceb7b53","Type":"ContainerDied","Data":"b19f1b5034590d1301f78db875436ea800cca53cc1b7466da8d58f24cfbaa75a"} Dec 10 19:09:21 crc kubenswrapper[4828]: I1210 19:09:21.660179 4828 generic.go:334] "Generic (PLEG): container finished" podID="288af64f-b8e1-4597-ab58-8fb61bf0d13c" containerID="3e54a0b3b54fd443dfeccf860b359b18ebe7d0bde4b93ac27f3b5e4ca13d5bb0" exitCode=0 Dec 10 19:09:21 crc kubenswrapper[4828]: I1210 19:09:21.660243 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grpgt" event={"ID":"288af64f-b8e1-4597-ab58-8fb61bf0d13c","Type":"ContainerDied","Data":"3e54a0b3b54fd443dfeccf860b359b18ebe7d0bde4b93ac27f3b5e4ca13d5bb0"} Dec 10 19:09:21 crc kubenswrapper[4828]: I1210 19:09:21.960928 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" Dec 10 19:09:21 crc kubenswrapper[4828]: I1210 19:09:21.965627 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.045871 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-bundle\") pod \"2c2bf8e7-87cf-4ede-bc49-e955290a66ee\" (UID: \"2c2bf8e7-87cf-4ede-bc49-e955290a66ee\") " Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.045917 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef8a4983-2aae-4e7e-a416-317ffceb7b53-util\") pod \"ef8a4983-2aae-4e7e-a416-317ffceb7b53\" (UID: \"ef8a4983-2aae-4e7e-a416-317ffceb7b53\") " Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.045968 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-util\") pod \"2c2bf8e7-87cf-4ede-bc49-e955290a66ee\" (UID: \"2c2bf8e7-87cf-4ede-bc49-e955290a66ee\") " Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.046011 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef8a4983-2aae-4e7e-a416-317ffceb7b53-bundle\") pod \"ef8a4983-2aae-4e7e-a416-317ffceb7b53\" (UID: \"ef8a4983-2aae-4e7e-a416-317ffceb7b53\") " Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.046065 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvc8q\" (UniqueName: \"kubernetes.io/projected/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-kube-api-access-qvc8q\") pod \"2c2bf8e7-87cf-4ede-bc49-e955290a66ee\" (UID: \"2c2bf8e7-87cf-4ede-bc49-e955290a66ee\") " Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.046112 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jf65\" (UniqueName: \"kubernetes.io/projected/ef8a4983-2aae-4e7e-a416-317ffceb7b53-kube-api-access-2jf65\") pod \"ef8a4983-2aae-4e7e-a416-317ffceb7b53\" (UID: \"ef8a4983-2aae-4e7e-a416-317ffceb7b53\") " Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.047494 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef8a4983-2aae-4e7e-a416-317ffceb7b53-bundle" (OuterVolumeSpecName: "bundle") pod "ef8a4983-2aae-4e7e-a416-317ffceb7b53" (UID: "ef8a4983-2aae-4e7e-a416-317ffceb7b53"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.048781 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-bundle" (OuterVolumeSpecName: "bundle") pod "2c2bf8e7-87cf-4ede-bc49-e955290a66ee" (UID: "2c2bf8e7-87cf-4ede-bc49-e955290a66ee"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.057835 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-kube-api-access-qvc8q" (OuterVolumeSpecName: "kube-api-access-qvc8q") pod "2c2bf8e7-87cf-4ede-bc49-e955290a66ee" (UID: "2c2bf8e7-87cf-4ede-bc49-e955290a66ee"). InnerVolumeSpecName "kube-api-access-qvc8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.058418 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef8a4983-2aae-4e7e-a416-317ffceb7b53-kube-api-access-2jf65" (OuterVolumeSpecName: "kube-api-access-2jf65") pod "ef8a4983-2aae-4e7e-a416-317ffceb7b53" (UID: "ef8a4983-2aae-4e7e-a416-317ffceb7b53"). InnerVolumeSpecName "kube-api-access-2jf65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.061886 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-util" (OuterVolumeSpecName: "util") pod "2c2bf8e7-87cf-4ede-bc49-e955290a66ee" (UID: "2c2bf8e7-87cf-4ede-bc49-e955290a66ee"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.070620 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef8a4983-2aae-4e7e-a416-317ffceb7b53-util" (OuterVolumeSpecName: "util") pod "ef8a4983-2aae-4e7e-a416-317ffceb7b53" (UID: "ef8a4983-2aae-4e7e-a416-317ffceb7b53"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.147394 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvc8q\" (UniqueName: \"kubernetes.io/projected/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-kube-api-access-qvc8q\") on node \"crc\" DevicePath \"\"" Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.147690 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jf65\" (UniqueName: \"kubernetes.io/projected/ef8a4983-2aae-4e7e-a416-317ffceb7b53-kube-api-access-2jf65\") on node \"crc\" DevicePath \"\"" Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.147700 4828 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.147711 4828 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ef8a4983-2aae-4e7e-a416-317ffceb7b53-util\") on node \"crc\" DevicePath \"\"" Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.147720 4828 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2c2bf8e7-87cf-4ede-bc49-e955290a66ee-util\") on node \"crc\" DevicePath \"\"" Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.147727 4828 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ef8a4983-2aae-4e7e-a416-317ffceb7b53-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.668902 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" event={"ID":"2c2bf8e7-87cf-4ede-bc49-e955290a66ee","Type":"ContainerDied","Data":"0045f5fc8870cca613d1a05398debe2d46c8b1fca11fc69dcc153fafeed0a67b"} Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.668941 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws" Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.668947 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0045f5fc8870cca613d1a05398debe2d46c8b1fca11fc69dcc153fafeed0a67b" Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.671055 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" event={"ID":"ef8a4983-2aae-4e7e-a416-317ffceb7b53","Type":"ContainerDied","Data":"89fe9acdaee267439be2d9fa74dfa7e96d6af84472f2802d41fa85fe0c1c8cf2"} Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.671078 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89fe9acdaee267439be2d9fa74dfa7e96d6af84472f2802d41fa85fe0c1c8cf2" Dec 10 19:09:22 crc kubenswrapper[4828]: I1210 19:09:22.671141 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5" Dec 10 19:09:24 crc kubenswrapper[4828]: I1210 19:09:24.684455 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grpgt" event={"ID":"288af64f-b8e1-4597-ab58-8fb61bf0d13c","Type":"ContainerStarted","Data":"a4fdfad3ad95c5a3e8b192ef0ca8e60795133b79a19b615e90c2f6759bccdccb"} Dec 10 19:09:24 crc kubenswrapper[4828]: I1210 19:09:24.712589 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-grpgt" podStartSLOduration=4.507402249 podStartE2EDuration="9.712564294s" podCreationTimestamp="2025-12-10 19:09:15 +0000 UTC" firstStartedPulling="2025-12-10 19:09:18.619023951 +0000 UTC m=+839.129634966" lastFinishedPulling="2025-12-10 19:09:23.824186006 +0000 UTC m=+844.334797011" observedRunningTime="2025-12-10 19:09:24.708840797 +0000 UTC m=+845.219451812" watchObservedRunningTime="2025-12-10 19:09:24.712564294 +0000 UTC m=+845.223175299" Dec 10 19:09:25 crc kubenswrapper[4828]: I1210 19:09:25.992236 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-grpgt" Dec 10 19:09:25 crc kubenswrapper[4828]: I1210 19:09:25.992536 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-grpgt" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.033845 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-grpgt" podUID="288af64f-b8e1-4597-ab58-8fb61bf0d13c" containerName="registry-server" probeResult="failure" output=< Dec 10 19:09:27 crc kubenswrapper[4828]: timeout: failed to connect service ":50051" within 1s Dec 10 19:09:27 crc kubenswrapper[4828]: > Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.537128 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42"] Dec 10 19:09:27 crc kubenswrapper[4828]: E1210 19:09:27.537605 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef8a4983-2aae-4e7e-a416-317ffceb7b53" containerName="extract" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.537619 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef8a4983-2aae-4e7e-a416-317ffceb7b53" containerName="extract" Dec 10 19:09:27 crc kubenswrapper[4828]: E1210 19:09:27.537629 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef8a4983-2aae-4e7e-a416-317ffceb7b53" containerName="util" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.537637 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef8a4983-2aae-4e7e-a416-317ffceb7b53" containerName="util" Dec 10 19:09:27 crc kubenswrapper[4828]: E1210 19:09:27.537645 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c2bf8e7-87cf-4ede-bc49-e955290a66ee" containerName="util" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.537652 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c2bf8e7-87cf-4ede-bc49-e955290a66ee" containerName="util" Dec 10 19:09:27 crc kubenswrapper[4828]: E1210 19:09:27.537667 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef8a4983-2aae-4e7e-a416-317ffceb7b53" containerName="pull" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.537672 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef8a4983-2aae-4e7e-a416-317ffceb7b53" containerName="pull" Dec 10 19:09:27 crc kubenswrapper[4828]: E1210 19:09:27.537679 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c2bf8e7-87cf-4ede-bc49-e955290a66ee" containerName="extract" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.537684 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c2bf8e7-87cf-4ede-bc49-e955290a66ee" containerName="extract" Dec 10 19:09:27 crc kubenswrapper[4828]: E1210 19:09:27.537695 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c2bf8e7-87cf-4ede-bc49-e955290a66ee" containerName="pull" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.537726 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c2bf8e7-87cf-4ede-bc49-e955290a66ee" containerName="pull" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.537908 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c2bf8e7-87cf-4ede-bc49-e955290a66ee" containerName="extract" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.537934 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef8a4983-2aae-4e7e-a416-317ffceb7b53" containerName="extract" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.538553 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.545131 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.552141 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-h4dpd" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.552403 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.552913 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.553080 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.555567 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.555977 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42"] Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.620180 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f4ae06be-8676-4704-aa09-b27912dd5dbf-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-9c7bf5466-2tc42\" (UID: \"f4ae06be-8676-4704-aa09-b27912dd5dbf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.620264 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f4ae06be-8676-4704-aa09-b27912dd5dbf-webhook-cert\") pod \"loki-operator-controller-manager-9c7bf5466-2tc42\" (UID: \"f4ae06be-8676-4704-aa09-b27912dd5dbf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.620300 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fzpw\" (UniqueName: \"kubernetes.io/projected/f4ae06be-8676-4704-aa09-b27912dd5dbf-kube-api-access-7fzpw\") pod \"loki-operator-controller-manager-9c7bf5466-2tc42\" (UID: \"f4ae06be-8676-4704-aa09-b27912dd5dbf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.620327 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/f4ae06be-8676-4704-aa09-b27912dd5dbf-manager-config\") pod \"loki-operator-controller-manager-9c7bf5466-2tc42\" (UID: \"f4ae06be-8676-4704-aa09-b27912dd5dbf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.620351 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f4ae06be-8676-4704-aa09-b27912dd5dbf-apiservice-cert\") pod \"loki-operator-controller-manager-9c7bf5466-2tc42\" (UID: \"f4ae06be-8676-4704-aa09-b27912dd5dbf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.721551 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f4ae06be-8676-4704-aa09-b27912dd5dbf-apiservice-cert\") pod \"loki-operator-controller-manager-9c7bf5466-2tc42\" (UID: \"f4ae06be-8676-4704-aa09-b27912dd5dbf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.721666 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f4ae06be-8676-4704-aa09-b27912dd5dbf-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-9c7bf5466-2tc42\" (UID: \"f4ae06be-8676-4704-aa09-b27912dd5dbf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.721750 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f4ae06be-8676-4704-aa09-b27912dd5dbf-webhook-cert\") pod \"loki-operator-controller-manager-9c7bf5466-2tc42\" (UID: \"f4ae06be-8676-4704-aa09-b27912dd5dbf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.721783 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fzpw\" (UniqueName: \"kubernetes.io/projected/f4ae06be-8676-4704-aa09-b27912dd5dbf-kube-api-access-7fzpw\") pod \"loki-operator-controller-manager-9c7bf5466-2tc42\" (UID: \"f4ae06be-8676-4704-aa09-b27912dd5dbf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.721824 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/f4ae06be-8676-4704-aa09-b27912dd5dbf-manager-config\") pod \"loki-operator-controller-manager-9c7bf5466-2tc42\" (UID: \"f4ae06be-8676-4704-aa09-b27912dd5dbf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.723133 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/f4ae06be-8676-4704-aa09-b27912dd5dbf-manager-config\") pod \"loki-operator-controller-manager-9c7bf5466-2tc42\" (UID: \"f4ae06be-8676-4704-aa09-b27912dd5dbf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.734195 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f4ae06be-8676-4704-aa09-b27912dd5dbf-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-9c7bf5466-2tc42\" (UID: \"f4ae06be-8676-4704-aa09-b27912dd5dbf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.737021 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f4ae06be-8676-4704-aa09-b27912dd5dbf-apiservice-cert\") pod \"loki-operator-controller-manager-9c7bf5466-2tc42\" (UID: \"f4ae06be-8676-4704-aa09-b27912dd5dbf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.746250 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f4ae06be-8676-4704-aa09-b27912dd5dbf-webhook-cert\") pod \"loki-operator-controller-manager-9c7bf5466-2tc42\" (UID: \"f4ae06be-8676-4704-aa09-b27912dd5dbf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.746465 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fzpw\" (UniqueName: \"kubernetes.io/projected/f4ae06be-8676-4704-aa09-b27912dd5dbf-kube-api-access-7fzpw\") pod \"loki-operator-controller-manager-9c7bf5466-2tc42\" (UID: \"f4ae06be-8676-4704-aa09-b27912dd5dbf\") " pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:27 crc kubenswrapper[4828]: I1210 19:09:27.856570 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:28 crc kubenswrapper[4828]: I1210 19:09:28.270783 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42"] Dec 10 19:09:28 crc kubenswrapper[4828]: W1210 19:09:28.271009 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4ae06be_8676_4704_aa09_b27912dd5dbf.slice/crio-92be070bce030308eaaaecd0b132c0a424e22d42cfac7c25ba264d2de79ae302 WatchSource:0}: Error finding container 92be070bce030308eaaaecd0b132c0a424e22d42cfac7c25ba264d2de79ae302: Status 404 returned error can't find the container with id 92be070bce030308eaaaecd0b132c0a424e22d42cfac7c25ba264d2de79ae302 Dec 10 19:09:28 crc kubenswrapper[4828]: I1210 19:09:28.708176 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" event={"ID":"f4ae06be-8676-4704-aa09-b27912dd5dbf","Type":"ContainerStarted","Data":"92be070bce030308eaaaecd0b132c0a424e22d42cfac7c25ba264d2de79ae302"} Dec 10 19:09:35 crc kubenswrapper[4828]: I1210 19:09:35.756559 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" event={"ID":"f4ae06be-8676-4704-aa09-b27912dd5dbf","Type":"ContainerStarted","Data":"5379bbe9bd77bfca30bf41beaf50407ae3f461527ea432bd076f6fafdb22393f"} Dec 10 19:09:36 crc kubenswrapper[4828]: I1210 19:09:36.049094 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-grpgt" Dec 10 19:09:36 crc kubenswrapper[4828]: I1210 19:09:36.089316 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-grpgt" Dec 10 19:09:37 crc kubenswrapper[4828]: I1210 19:09:37.798892 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-b94hr"] Dec 10 19:09:37 crc kubenswrapper[4828]: I1210 19:09:37.799937 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-b94hr" Dec 10 19:09:37 crc kubenswrapper[4828]: I1210 19:09:37.808174 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Dec 10 19:09:37 crc kubenswrapper[4828]: I1210 19:09:37.808255 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Dec 10 19:09:37 crc kubenswrapper[4828]: I1210 19:09:37.808360 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-q9czs" Dec 10 19:09:37 crc kubenswrapper[4828]: I1210 19:09:37.838338 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-b94hr"] Dec 10 19:09:37 crc kubenswrapper[4828]: I1210 19:09:37.889458 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjcrm\" (UniqueName: \"kubernetes.io/projected/fe41e1e4-96fb-4798-9dc8-3382d8c74b46-kube-api-access-hjcrm\") pod \"cluster-logging-operator-ff9846bd-b94hr\" (UID: \"fe41e1e4-96fb-4798-9dc8-3382d8c74b46\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-b94hr" Dec 10 19:09:37 crc kubenswrapper[4828]: I1210 19:09:37.990872 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjcrm\" (UniqueName: \"kubernetes.io/projected/fe41e1e4-96fb-4798-9dc8-3382d8c74b46-kube-api-access-hjcrm\") pod \"cluster-logging-operator-ff9846bd-b94hr\" (UID: \"fe41e1e4-96fb-4798-9dc8-3382d8c74b46\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-b94hr" Dec 10 19:09:38 crc kubenswrapper[4828]: I1210 19:09:38.030648 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjcrm\" (UniqueName: \"kubernetes.io/projected/fe41e1e4-96fb-4798-9dc8-3382d8c74b46-kube-api-access-hjcrm\") pod \"cluster-logging-operator-ff9846bd-b94hr\" (UID: \"fe41e1e4-96fb-4798-9dc8-3382d8c74b46\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-b94hr" Dec 10 19:09:38 crc kubenswrapper[4828]: I1210 19:09:38.192607 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-b94hr" Dec 10 19:09:39 crc kubenswrapper[4828]: I1210 19:09:39.658432 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-grpgt"] Dec 10 19:09:39 crc kubenswrapper[4828]: I1210 19:09:39.658941 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-grpgt" podUID="288af64f-b8e1-4597-ab58-8fb61bf0d13c" containerName="registry-server" containerID="cri-o://a4fdfad3ad95c5a3e8b192ef0ca8e60795133b79a19b615e90c2f6759bccdccb" gracePeriod=2 Dec 10 19:09:40 crc kubenswrapper[4828]: I1210 19:09:40.812498 4828 generic.go:334] "Generic (PLEG): container finished" podID="288af64f-b8e1-4597-ab58-8fb61bf0d13c" containerID="a4fdfad3ad95c5a3e8b192ef0ca8e60795133b79a19b615e90c2f6759bccdccb" exitCode=0 Dec 10 19:09:40 crc kubenswrapper[4828]: I1210 19:09:40.812705 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grpgt" event={"ID":"288af64f-b8e1-4597-ab58-8fb61bf0d13c","Type":"ContainerDied","Data":"a4fdfad3ad95c5a3e8b192ef0ca8e60795133b79a19b615e90c2f6759bccdccb"} Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.297889 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-grpgt" Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.452776 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrb7k\" (UniqueName: \"kubernetes.io/projected/288af64f-b8e1-4597-ab58-8fb61bf0d13c-kube-api-access-jrb7k\") pod \"288af64f-b8e1-4597-ab58-8fb61bf0d13c\" (UID: \"288af64f-b8e1-4597-ab58-8fb61bf0d13c\") " Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.453399 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/288af64f-b8e1-4597-ab58-8fb61bf0d13c-catalog-content\") pod \"288af64f-b8e1-4597-ab58-8fb61bf0d13c\" (UID: \"288af64f-b8e1-4597-ab58-8fb61bf0d13c\") " Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.453922 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/288af64f-b8e1-4597-ab58-8fb61bf0d13c-utilities\") pod \"288af64f-b8e1-4597-ab58-8fb61bf0d13c\" (UID: \"288af64f-b8e1-4597-ab58-8fb61bf0d13c\") " Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.454673 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/288af64f-b8e1-4597-ab58-8fb61bf0d13c-utilities" (OuterVolumeSpecName: "utilities") pod "288af64f-b8e1-4597-ab58-8fb61bf0d13c" (UID: "288af64f-b8e1-4597-ab58-8fb61bf0d13c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.459004 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/288af64f-b8e1-4597-ab58-8fb61bf0d13c-kube-api-access-jrb7k" (OuterVolumeSpecName: "kube-api-access-jrb7k") pod "288af64f-b8e1-4597-ab58-8fb61bf0d13c" (UID: "288af64f-b8e1-4597-ab58-8fb61bf0d13c"). InnerVolumeSpecName "kube-api-access-jrb7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.555493 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/288af64f-b8e1-4597-ab58-8fb61bf0d13c-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.555534 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrb7k\" (UniqueName: \"kubernetes.io/projected/288af64f-b8e1-4597-ab58-8fb61bf0d13c-kube-api-access-jrb7k\") on node \"crc\" DevicePath \"\"" Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.558486 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/288af64f-b8e1-4597-ab58-8fb61bf0d13c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "288af64f-b8e1-4597-ab58-8fb61bf0d13c" (UID: "288af64f-b8e1-4597-ab58-8fb61bf0d13c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.629433 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-b94hr"] Dec 10 19:09:41 crc kubenswrapper[4828]: W1210 19:09:41.635249 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe41e1e4_96fb_4798_9dc8_3382d8c74b46.slice/crio-b0e6f7dce84a0f903c33cce7ed9abb33a17f6ab13c1fb76f7e4df3ef8d8777db WatchSource:0}: Error finding container b0e6f7dce84a0f903c33cce7ed9abb33a17f6ab13c1fb76f7e4df3ef8d8777db: Status 404 returned error can't find the container with id b0e6f7dce84a0f903c33cce7ed9abb33a17f6ab13c1fb76f7e4df3ef8d8777db Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.656839 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/288af64f-b8e1-4597-ab58-8fb61bf0d13c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.824162 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-grpgt" event={"ID":"288af64f-b8e1-4597-ab58-8fb61bf0d13c","Type":"ContainerDied","Data":"f75547a606f98737a864ab17e97987385e6773f62b8b352165a0144864f170b0"} Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.824204 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-grpgt" Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.824212 4828 scope.go:117] "RemoveContainer" containerID="a4fdfad3ad95c5a3e8b192ef0ca8e60795133b79a19b615e90c2f6759bccdccb" Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.826024 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-b94hr" event={"ID":"fe41e1e4-96fb-4798-9dc8-3382d8c74b46","Type":"ContainerStarted","Data":"b0e6f7dce84a0f903c33cce7ed9abb33a17f6ab13c1fb76f7e4df3ef8d8777db"} Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.831741 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" event={"ID":"f4ae06be-8676-4704-aa09-b27912dd5dbf","Type":"ContainerStarted","Data":"02e42a2eca35bf69e8d2af817e77140673c889160b62d949d793090c5775a384"} Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.832784 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.840659 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.845846 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-grpgt"] Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.847451 4828 scope.go:117] "RemoveContainer" containerID="3e54a0b3b54fd443dfeccf860b359b18ebe7d0bde4b93ac27f3b5e4ca13d5bb0" Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.849854 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-grpgt"] Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.872070 4828 scope.go:117] "RemoveContainer" containerID="61b0dd212618f1bbf2d7229a16df935f3b48abce99b3f5504c6d47cd34e10608" Dec 10 19:09:41 crc kubenswrapper[4828]: I1210 19:09:41.873014 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-9c7bf5466-2tc42" podStartSLOduration=1.822182424 podStartE2EDuration="14.873005791s" podCreationTimestamp="2025-12-10 19:09:27 +0000 UTC" firstStartedPulling="2025-12-10 19:09:28.273958293 +0000 UTC m=+848.784569298" lastFinishedPulling="2025-12-10 19:09:41.32478166 +0000 UTC m=+861.835392665" observedRunningTime="2025-12-10 19:09:41.87258629 +0000 UTC m=+862.383197315" watchObservedRunningTime="2025-12-10 19:09:41.873005791 +0000 UTC m=+862.383616796" Dec 10 19:09:43 crc kubenswrapper[4828]: I1210 19:09:43.805539 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="288af64f-b8e1-4597-ab58-8fb61bf0d13c" path="/var/lib/kubelet/pods/288af64f-b8e1-4597-ab58-8fb61bf0d13c/volumes" Dec 10 19:09:46 crc kubenswrapper[4828]: I1210 19:09:46.883277 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-b94hr" event={"ID":"fe41e1e4-96fb-4798-9dc8-3382d8c74b46","Type":"ContainerStarted","Data":"c130e5e9099bcb76b0f1481ffd0317120e948f6cc60670987623ce90ccc1ca4d"} Dec 10 19:09:46 crc kubenswrapper[4828]: I1210 19:09:46.900790 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-ff9846bd-b94hr" podStartSLOduration=4.826392344 podStartE2EDuration="9.900771641s" podCreationTimestamp="2025-12-10 19:09:37 +0000 UTC" firstStartedPulling="2025-12-10 19:09:41.637480326 +0000 UTC m=+862.148091331" lastFinishedPulling="2025-12-10 19:09:46.711859623 +0000 UTC m=+867.222470628" observedRunningTime="2025-12-10 19:09:46.899298612 +0000 UTC m=+867.409909627" watchObservedRunningTime="2025-12-10 19:09:46.900771641 +0000 UTC m=+867.411382646" Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.681211 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Dec 10 19:09:51 crc kubenswrapper[4828]: E1210 19:09:51.682001 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="288af64f-b8e1-4597-ab58-8fb61bf0d13c" containerName="extract-content" Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.682018 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="288af64f-b8e1-4597-ab58-8fb61bf0d13c" containerName="extract-content" Dec 10 19:09:51 crc kubenswrapper[4828]: E1210 19:09:51.682038 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="288af64f-b8e1-4597-ab58-8fb61bf0d13c" containerName="extract-utilities" Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.682046 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="288af64f-b8e1-4597-ab58-8fb61bf0d13c" containerName="extract-utilities" Dec 10 19:09:51 crc kubenswrapper[4828]: E1210 19:09:51.682059 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="288af64f-b8e1-4597-ab58-8fb61bf0d13c" containerName="registry-server" Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.682066 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="288af64f-b8e1-4597-ab58-8fb61bf0d13c" containerName="registry-server" Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.682220 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="288af64f-b8e1-4597-ab58-8fb61bf0d13c" containerName="registry-server" Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.682747 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.685236 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.685483 4828 reflector.go:368] Caches populated for *v1.Secret from object-"minio-dev"/"default-dockercfg-dfwzw" Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.685997 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.693086 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.816210 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58bpl\" (UniqueName: \"kubernetes.io/projected/d7f2b9bd-fe2e-4dc9-b16f-901b8422acd1-kube-api-access-58bpl\") pod \"minio\" (UID: \"d7f2b9bd-fe2e-4dc9-b16f-901b8422acd1\") " pod="minio-dev/minio" Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.816316 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2246da1d-ddf7-4cbf-8cf1-2dc98457ce84\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2246da1d-ddf7-4cbf-8cf1-2dc98457ce84\") pod \"minio\" (UID: \"d7f2b9bd-fe2e-4dc9-b16f-901b8422acd1\") " pod="minio-dev/minio" Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.917047 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2246da1d-ddf7-4cbf-8cf1-2dc98457ce84\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2246da1d-ddf7-4cbf-8cf1-2dc98457ce84\") pod \"minio\" (UID: \"d7f2b9bd-fe2e-4dc9-b16f-901b8422acd1\") " pod="minio-dev/minio" Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.917128 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58bpl\" (UniqueName: \"kubernetes.io/projected/d7f2b9bd-fe2e-4dc9-b16f-901b8422acd1-kube-api-access-58bpl\") pod \"minio\" (UID: \"d7f2b9bd-fe2e-4dc9-b16f-901b8422acd1\") " pod="minio-dev/minio" Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.922401 4828 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.922446 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2246da1d-ddf7-4cbf-8cf1-2dc98457ce84\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2246da1d-ddf7-4cbf-8cf1-2dc98457ce84\") pod \"minio\" (UID: \"d7f2b9bd-fe2e-4dc9-b16f-901b8422acd1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/573a1320f926126b5fdba2fc342da58c8df8d401674869d5f1c01837c703ae0a/globalmount\"" pod="minio-dev/minio" Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.939266 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58bpl\" (UniqueName: \"kubernetes.io/projected/d7f2b9bd-fe2e-4dc9-b16f-901b8422acd1-kube-api-access-58bpl\") pod \"minio\" (UID: \"d7f2b9bd-fe2e-4dc9-b16f-901b8422acd1\") " pod="minio-dev/minio" Dec 10 19:09:51 crc kubenswrapper[4828]: I1210 19:09:51.943570 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2246da1d-ddf7-4cbf-8cf1-2dc98457ce84\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2246da1d-ddf7-4cbf-8cf1-2dc98457ce84\") pod \"minio\" (UID: \"d7f2b9bd-fe2e-4dc9-b16f-901b8422acd1\") " pod="minio-dev/minio" Dec 10 19:09:52 crc kubenswrapper[4828]: I1210 19:09:52.002487 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 10 19:09:52 crc kubenswrapper[4828]: I1210 19:09:52.529177 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 10 19:09:52 crc kubenswrapper[4828]: I1210 19:09:52.928594 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"d7f2b9bd-fe2e-4dc9-b16f-901b8422acd1","Type":"ContainerStarted","Data":"16da6611d345eed09e8d8ade668c10a12489eedc13ff93ad44f681a27f25687a"} Dec 10 19:09:56 crc kubenswrapper[4828]: I1210 19:09:56.951596 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"d7f2b9bd-fe2e-4dc9-b16f-901b8422acd1","Type":"ContainerStarted","Data":"4baa844811c3bb9b9d911eb135ba89ccfb95f27c2c4c56975b7ba4480fb8ab31"} Dec 10 19:09:56 crc kubenswrapper[4828]: I1210 19:09:56.972616 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.617301849 podStartE2EDuration="7.972590355s" podCreationTimestamp="2025-12-10 19:09:49 +0000 UTC" firstStartedPulling="2025-12-10 19:09:52.53330607 +0000 UTC m=+873.043917095" lastFinishedPulling="2025-12-10 19:09:55.888594596 +0000 UTC m=+876.399205601" observedRunningTime="2025-12-10 19:09:56.964730021 +0000 UTC m=+877.475341096" watchObservedRunningTime="2025-12-10 19:09:56.972590355 +0000 UTC m=+877.483201400" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.633476 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-lg62n"] Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.635222 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.637681 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.637907 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.638047 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-h62g2" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.638153 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.639309 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.652608 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-lg62n"] Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.748908 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/14db470d-4202-4b2d-9339-65ba24e521d0-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-lg62n\" (UID: \"14db470d-4202-4b2d-9339-65ba24e521d0\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.748992 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14db470d-4202-4b2d-9339-65ba24e521d0-config\") pod \"logging-loki-distributor-76cc67bf56-lg62n\" (UID: \"14db470d-4202-4b2d-9339-65ba24e521d0\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.749055 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k66pj\" (UniqueName: \"kubernetes.io/projected/14db470d-4202-4b2d-9339-65ba24e521d0-kube-api-access-k66pj\") pod \"logging-loki-distributor-76cc67bf56-lg62n\" (UID: \"14db470d-4202-4b2d-9339-65ba24e521d0\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.749085 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14db470d-4202-4b2d-9339-65ba24e521d0-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-lg62n\" (UID: \"14db470d-4202-4b2d-9339-65ba24e521d0\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.749120 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/14db470d-4202-4b2d-9339-65ba24e521d0-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-lg62n\" (UID: \"14db470d-4202-4b2d-9339-65ba24e521d0\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.809469 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-4pkdl"] Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.810563 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.815324 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.815460 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.815520 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.830850 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-4pkdl"] Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.850792 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14db470d-4202-4b2d-9339-65ba24e521d0-config\") pod \"logging-loki-distributor-76cc67bf56-lg62n\" (UID: \"14db470d-4202-4b2d-9339-65ba24e521d0\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.850888 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k66pj\" (UniqueName: \"kubernetes.io/projected/14db470d-4202-4b2d-9339-65ba24e521d0-kube-api-access-k66pj\") pod \"logging-loki-distributor-76cc67bf56-lg62n\" (UID: \"14db470d-4202-4b2d-9339-65ba24e521d0\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.850915 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14db470d-4202-4b2d-9339-65ba24e521d0-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-lg62n\" (UID: \"14db470d-4202-4b2d-9339-65ba24e521d0\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.850940 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/14db470d-4202-4b2d-9339-65ba24e521d0-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-lg62n\" (UID: \"14db470d-4202-4b2d-9339-65ba24e521d0\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.851057 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/14db470d-4202-4b2d-9339-65ba24e521d0-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-lg62n\" (UID: \"14db470d-4202-4b2d-9339-65ba24e521d0\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.852523 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14db470d-4202-4b2d-9339-65ba24e521d0-config\") pod \"logging-loki-distributor-76cc67bf56-lg62n\" (UID: \"14db470d-4202-4b2d-9339-65ba24e521d0\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.853447 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/14db470d-4202-4b2d-9339-65ba24e521d0-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-lg62n\" (UID: \"14db470d-4202-4b2d-9339-65ba24e521d0\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.865910 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/14db470d-4202-4b2d-9339-65ba24e521d0-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-lg62n\" (UID: \"14db470d-4202-4b2d-9339-65ba24e521d0\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.871606 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/14db470d-4202-4b2d-9339-65ba24e521d0-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-lg62n\" (UID: \"14db470d-4202-4b2d-9339-65ba24e521d0\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.879872 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k66pj\" (UniqueName: \"kubernetes.io/projected/14db470d-4202-4b2d-9339-65ba24e521d0-kube-api-access-k66pj\") pod \"logging-loki-distributor-76cc67bf56-lg62n\" (UID: \"14db470d-4202-4b2d-9339-65ba24e521d0\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.899692 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb"] Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.900499 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.904097 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.910214 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.934520 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb"] Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.952045 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/15695070-e7cd-4af1-980d-e79defa1efaf-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.952148 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15695070-e7cd-4af1-980d-e79defa1efaf-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.952202 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zc6zm\" (UniqueName: \"kubernetes.io/projected/15695070-e7cd-4af1-980d-e79defa1efaf-kube-api-access-zc6zm\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.952232 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15695070-e7cd-4af1-980d-e79defa1efaf-config\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.952410 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/15695070-e7cd-4af1-980d-e79defa1efaf-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.952527 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/15695070-e7cd-4af1-980d-e79defa1efaf-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:00 crc kubenswrapper[4828]: I1210 19:10:00.955267 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.007368 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-6c4d974455-7r2h2"] Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.008742 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.013521 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.013860 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.013855 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-6c4d974455-7lpjj"] Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.014027 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.014177 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.014314 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.015300 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.017917 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-6c4d974455-7r2h2"] Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.017953 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-hfgnm" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.032751 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-6c4d974455-7lpjj"] Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.053277 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15695070-e7cd-4af1-980d-e79defa1efaf-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.053321 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cf6784a-0699-4d1c-ba27-cb8f25ab729b-config\") pod \"logging-loki-query-frontend-84558f7c9f-qqdqb\" (UID: \"6cf6784a-0699-4d1c-ba27-cb8f25ab729b\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.053349 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6cf6784a-0699-4d1c-ba27-cb8f25ab729b-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-qqdqb\" (UID: \"6cf6784a-0699-4d1c-ba27-cb8f25ab729b\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.053366 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zc6zm\" (UniqueName: \"kubernetes.io/projected/15695070-e7cd-4af1-980d-e79defa1efaf-kube-api-access-zc6zm\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.053383 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/6cf6784a-0699-4d1c-ba27-cb8f25ab729b-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-qqdqb\" (UID: \"6cf6784a-0699-4d1c-ba27-cb8f25ab729b\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.053404 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15695070-e7cd-4af1-980d-e79defa1efaf-config\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.053420 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/6cf6784a-0699-4d1c-ba27-cb8f25ab729b-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-qqdqb\" (UID: \"6cf6784a-0699-4d1c-ba27-cb8f25ab729b\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.053454 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/15695070-e7cd-4af1-980d-e79defa1efaf-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.053484 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/15695070-e7cd-4af1-980d-e79defa1efaf-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.053513 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/15695070-e7cd-4af1-980d-e79defa1efaf-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.053540 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlgjp\" (UniqueName: \"kubernetes.io/projected/6cf6784a-0699-4d1c-ba27-cb8f25ab729b-kube-api-access-wlgjp\") pod \"logging-loki-query-frontend-84558f7c9f-qqdqb\" (UID: \"6cf6784a-0699-4d1c-ba27-cb8f25ab729b\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.055158 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15695070-e7cd-4af1-980d-e79defa1efaf-config\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.056364 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15695070-e7cd-4af1-980d-e79defa1efaf-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.061147 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/15695070-e7cd-4af1-980d-e79defa1efaf-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.070313 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/15695070-e7cd-4af1-980d-e79defa1efaf-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.073232 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/15695070-e7cd-4af1-980d-e79defa1efaf-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.079591 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zc6zm\" (UniqueName: \"kubernetes.io/projected/15695070-e7cd-4af1-980d-e79defa1efaf-kube-api-access-zc6zm\") pod \"logging-loki-querier-5895d59bb8-4pkdl\" (UID: \"15695070-e7cd-4af1-980d-e79defa1efaf\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.124689 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.154455 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.154506 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/cf17217c-0dc6-4edd-8334-8896f103f9f1-rbac\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.154530 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-lokistack-gateway\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.154808 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj5qz\" (UniqueName: \"kubernetes.io/projected/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-kube-api-access-mj5qz\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.155000 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/cf17217c-0dc6-4edd-8334-8896f103f9f1-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.155079 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-tls-secret\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.155126 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlgjp\" (UniqueName: \"kubernetes.io/projected/6cf6784a-0699-4d1c-ba27-cb8f25ab729b-kube-api-access-wlgjp\") pod \"logging-loki-query-frontend-84558f7c9f-qqdqb\" (UID: \"6cf6784a-0699-4d1c-ba27-cb8f25ab729b\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.155151 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-tenants\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.155174 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/cf17217c-0dc6-4edd-8334-8896f103f9f1-tenants\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.155267 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cf6784a-0699-4d1c-ba27-cb8f25ab729b-config\") pod \"logging-loki-query-frontend-84558f7c9f-qqdqb\" (UID: \"6cf6784a-0699-4d1c-ba27-cb8f25ab729b\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.155295 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-rbac\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.155352 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6cf6784a-0699-4d1c-ba27-cb8f25ab729b-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-qqdqb\" (UID: \"6cf6784a-0699-4d1c-ba27-cb8f25ab729b\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.155384 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/6cf6784a-0699-4d1c-ba27-cb8f25ab729b-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-qqdqb\" (UID: \"6cf6784a-0699-4d1c-ba27-cb8f25ab729b\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.155422 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/6cf6784a-0699-4d1c-ba27-cb8f25ab729b-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-qqdqb\" (UID: \"6cf6784a-0699-4d1c-ba27-cb8f25ab729b\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.156786 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/cf17217c-0dc6-4edd-8334-8896f103f9f1-lokistack-gateway\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.156833 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h4fw\" (UniqueName: \"kubernetes.io/projected/cf17217c-0dc6-4edd-8334-8896f103f9f1-kube-api-access-9h4fw\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.156870 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf17217c-0dc6-4edd-8334-8896f103f9f1-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.156898 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/cf17217c-0dc6-4edd-8334-8896f103f9f1-tls-secret\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.156919 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf17217c-0dc6-4edd-8334-8896f103f9f1-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.156952 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.156997 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.158168 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cf6784a-0699-4d1c-ba27-cb8f25ab729b-config\") pod \"logging-loki-query-frontend-84558f7c9f-qqdqb\" (UID: \"6cf6784a-0699-4d1c-ba27-cb8f25ab729b\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.159847 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6cf6784a-0699-4d1c-ba27-cb8f25ab729b-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-qqdqb\" (UID: \"6cf6784a-0699-4d1c-ba27-cb8f25ab729b\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.164470 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/6cf6784a-0699-4d1c-ba27-cb8f25ab729b-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-qqdqb\" (UID: \"6cf6784a-0699-4d1c-ba27-cb8f25ab729b\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.169140 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/6cf6784a-0699-4d1c-ba27-cb8f25ab729b-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-qqdqb\" (UID: \"6cf6784a-0699-4d1c-ba27-cb8f25ab729b\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.179090 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlgjp\" (UniqueName: \"kubernetes.io/projected/6cf6784a-0699-4d1c-ba27-cb8f25ab729b-kube-api-access-wlgjp\") pod \"logging-loki-query-frontend-84558f7c9f-qqdqb\" (UID: \"6cf6784a-0699-4d1c-ba27-cb8f25ab729b\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.228910 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.259555 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.259609 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.259636 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.259653 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/cf17217c-0dc6-4edd-8334-8896f103f9f1-rbac\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.259669 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-lokistack-gateway\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.259689 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj5qz\" (UniqueName: \"kubernetes.io/projected/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-kube-api-access-mj5qz\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.259708 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/cf17217c-0dc6-4edd-8334-8896f103f9f1-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.259730 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-tls-secret\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.259746 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-tenants\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.259759 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/cf17217c-0dc6-4edd-8334-8896f103f9f1-tenants\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.259810 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-rbac\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.259849 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/cf17217c-0dc6-4edd-8334-8896f103f9f1-lokistack-gateway\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.259865 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h4fw\" (UniqueName: \"kubernetes.io/projected/cf17217c-0dc6-4edd-8334-8896f103f9f1-kube-api-access-9h4fw\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.259886 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf17217c-0dc6-4edd-8334-8896f103f9f1-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.259903 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/cf17217c-0dc6-4edd-8334-8896f103f9f1-tls-secret\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.259922 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf17217c-0dc6-4edd-8334-8896f103f9f1-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.260622 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf17217c-0dc6-4edd-8334-8896f103f9f1-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.261299 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.262940 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-rbac\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.263720 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/cf17217c-0dc6-4edd-8334-8896f103f9f1-rbac\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.263945 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/cf17217c-0dc6-4edd-8334-8896f103f9f1-lokistack-gateway\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.264814 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-tls-secret\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.264899 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.265026 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf17217c-0dc6-4edd-8334-8896f103f9f1-logging-loki-ca-bundle\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.266177 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-tenants\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.266643 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/cf17217c-0dc6-4edd-8334-8896f103f9f1-tenants\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.267164 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/cf17217c-0dc6-4edd-8334-8896f103f9f1-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.268190 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/cf17217c-0dc6-4edd-8334-8896f103f9f1-tls-secret\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.268418 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-lokistack-gateway\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.269471 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.283404 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h4fw\" (UniqueName: \"kubernetes.io/projected/cf17217c-0dc6-4edd-8334-8896f103f9f1-kube-api-access-9h4fw\") pod \"logging-loki-gateway-6c4d974455-7lpjj\" (UID: \"cf17217c-0dc6-4edd-8334-8896f103f9f1\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.291881 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj5qz\" (UniqueName: \"kubernetes.io/projected/4c744ad4-db8c-42ad-9b49-ecb32b023cb6-kube-api-access-mj5qz\") pod \"logging-loki-gateway-6c4d974455-7r2h2\" (UID: \"4c744ad4-db8c-42ad-9b49-ecb32b023cb6\") " pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.355583 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.382417 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.453211 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-lg62n"] Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.576088 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-4pkdl"] Dec 10 19:10:01 crc kubenswrapper[4828]: W1210 19:10:01.592530 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15695070_e7cd_4af1_980d_e79defa1efaf.slice/crio-f08c9d82c692fdb3958338e6066949a57634f108d6596f9db6433f622d022e5f WatchSource:0}: Error finding container f08c9d82c692fdb3958338e6066949a57634f108d6596f9db6433f622d022e5f: Status 404 returned error can't find the container with id f08c9d82c692fdb3958338e6066949a57634f108d6596f9db6433f622d022e5f Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.680929 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb"] Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.798249 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.800641 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.803762 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.808671 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.808824 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Dec 10 19:10:01 crc kubenswrapper[4828]: W1210 19:10:01.849472 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c744ad4_db8c_42ad_9b49_ecb32b023cb6.slice/crio-b87d13095971866a7a250717fddd10679b7ed903bc054b60a18d9217ec6b65d8 WatchSource:0}: Error finding container b87d13095971866a7a250717fddd10679b7ed903bc054b60a18d9217ec6b65d8: Status 404 returned error can't find the container with id b87d13095971866a7a250717fddd10679b7ed903bc054b60a18d9217ec6b65d8 Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.858396 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-6c4d974455-7r2h2"] Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.885615 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/a4483694-bd44-4582-80b0-17c774d39682-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.885674 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a4483694-bd44-4582-80b0-17c774d39682-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.885700 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms82v\" (UniqueName: \"kubernetes.io/projected/a4483694-bd44-4582-80b0-17c774d39682-kube-api-access-ms82v\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.885754 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/a4483694-bd44-4582-80b0-17c774d39682-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.885915 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-aa2949b6-65f7-4a1c-b058-21e17b10693b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-aa2949b6-65f7-4a1c-b058-21e17b10693b\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.885938 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4483694-bd44-4582-80b0-17c774d39682-config\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.885982 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-78db4c05-3a69-4cf4-a0a9-1f14bd2919b6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-78db4c05-3a69-4cf4-a0a9-1f14bd2919b6\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.886005 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/a4483694-bd44-4582-80b0-17c774d39682-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: W1210 19:10:01.898021 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf17217c_0dc6_4edd_8334_8896f103f9f1.slice/crio-fbea50c22527992859d3964df867877a821d8a8a340312a06d92f217019c99fd WatchSource:0}: Error finding container fbea50c22527992859d3964df867877a821d8a8a340312a06d92f217019c99fd: Status 404 returned error can't find the container with id fbea50c22527992859d3964df867877a821d8a8a340312a06d92f217019c99fd Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.907087 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-6c4d974455-7lpjj"] Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.921228 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.922473 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.924553 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.925152 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.925339 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.957553 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.958552 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.962285 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.962527 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.968162 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.987030 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" event={"ID":"15695070-e7cd-4af1-980d-e79defa1efaf","Type":"ContainerStarted","Data":"f08c9d82c692fdb3958338e6066949a57634f108d6596f9db6433f622d022e5f"} Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.988631 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4483694-bd44-4582-80b0-17c774d39682-config\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.988774 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" event={"ID":"14db470d-4202-4b2d-9339-65ba24e521d0","Type":"ContainerStarted","Data":"50765d9b3c7394d9ca5118def9561237e9d3ac8c56a3766e09bd56006d1c71d4"} Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.989262 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-78db4c05-3a69-4cf4-a0a9-1f14bd2919b6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-78db4c05-3a69-4cf4-a0a9-1f14bd2919b6\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.989309 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3501ce8-e4b7-420d-b727-5ed3015f39d2-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.989335 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/d3501ce8-e4b7-420d-b727-5ed3015f39d2-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.989375 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/a4483694-bd44-4582-80b0-17c774d39682-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.989523 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d81fdf23-e471-4a74-a4ac-24e665108494\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d81fdf23-e471-4a74-a4ac-24e665108494\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.989577 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/a4483694-bd44-4582-80b0-17c774d39682-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.989609 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a4483694-bd44-4582-80b0-17c774d39682-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.989636 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms82v\" (UniqueName: \"kubernetes.io/projected/a4483694-bd44-4582-80b0-17c774d39682-kube-api-access-ms82v\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.989865 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkqkn\" (UniqueName: \"kubernetes.io/projected/d3501ce8-e4b7-420d-b727-5ed3015f39d2-kube-api-access-mkqkn\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.989941 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/d3501ce8-e4b7-420d-b727-5ed3015f39d2-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.990005 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/a4483694-bd44-4582-80b0-17c774d39682-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.990035 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3501ce8-e4b7-420d-b727-5ed3015f39d2-config\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.990066 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/d3501ce8-e4b7-420d-b727-5ed3015f39d2-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.990121 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-aa2949b6-65f7-4a1c-b058-21e17b10693b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-aa2949b6-65f7-4a1c-b058-21e17b10693b\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.990274 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" event={"ID":"6cf6784a-0699-4d1c-ba27-cb8f25ab729b","Type":"ContainerStarted","Data":"d58fff8bffa1bbf382b7c8b3575742b201e5bf34b2e92d0488f4182527f3372e"} Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.991388 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a4483694-bd44-4582-80b0-17c774d39682-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.991464 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4483694-bd44-4582-80b0-17c774d39682-config\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.992541 4828 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.992578 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-78db4c05-3a69-4cf4-a0a9-1f14bd2919b6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-78db4c05-3a69-4cf4-a0a9-1f14bd2919b6\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2036e95e69c587e5fc07fab25fddbc6ebb0b236cf74106b1f1232a7f92f1a44e/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.993466 4828 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.993501 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-aa2949b6-65f7-4a1c-b058-21e17b10693b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-aa2949b6-65f7-4a1c-b058-21e17b10693b\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/fd4446a6f30a765cc688f58e5f296b13f0b757fe8f48912b305ef33c25b97c3c/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.994720 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" event={"ID":"4c744ad4-db8c-42ad-9b49-ecb32b023cb6","Type":"ContainerStarted","Data":"b87d13095971866a7a250717fddd10679b7ed903bc054b60a18d9217ec6b65d8"} Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.995029 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/a4483694-bd44-4582-80b0-17c774d39682-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.996332 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" event={"ID":"cf17217c-0dc6-4edd-8334-8896f103f9f1","Type":"ContainerStarted","Data":"fbea50c22527992859d3964df867877a821d8a8a340312a06d92f217019c99fd"} Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.996441 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/a4483694-bd44-4582-80b0-17c774d39682-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:01 crc kubenswrapper[4828]: I1210 19:10:01.998048 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/a4483694-bd44-4582-80b0-17c774d39682-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.005486 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms82v\" (UniqueName: \"kubernetes.io/projected/a4483694-bd44-4582-80b0-17c774d39682-kube-api-access-ms82v\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.022116 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-78db4c05-3a69-4cf4-a0a9-1f14bd2919b6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-78db4c05-3a69-4cf4-a0a9-1f14bd2919b6\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.023211 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-aa2949b6-65f7-4a1c-b058-21e17b10693b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-aa2949b6-65f7-4a1c-b058-21e17b10693b\") pod \"logging-loki-ingester-0\" (UID: \"a4483694-bd44-4582-80b0-17c774d39682\") " pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.091491 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df650a21-47af-445a-9b8a-8908684aa6b9-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.091815 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d81fdf23-e471-4a74-a4ac-24e665108494\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d81fdf23-e471-4a74-a4ac-24e665108494\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.091927 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/d3501ce8-e4b7-420d-b727-5ed3015f39d2-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.092056 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxzps\" (UniqueName: \"kubernetes.io/projected/df650a21-47af-445a-9b8a-8908684aa6b9-kube-api-access-hxzps\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.092305 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/df650a21-47af-445a-9b8a-8908684aa6b9-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.092410 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3501ce8-e4b7-420d-b727-5ed3015f39d2-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.092976 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/d3501ce8-e4b7-420d-b727-5ed3015f39d2-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.093083 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/df650a21-47af-445a-9b8a-8908684aa6b9-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.093295 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkqkn\" (UniqueName: \"kubernetes.io/projected/d3501ce8-e4b7-420d-b727-5ed3015f39d2-kube-api-access-mkqkn\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.093366 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3501ce8-e4b7-420d-b727-5ed3015f39d2-config\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.093418 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/d3501ce8-e4b7-420d-b727-5ed3015f39d2-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.093515 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d3422407-00c4-4e72-96b2-cb91ea9c78e8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d3422407-00c4-4e72-96b2-cb91ea9c78e8\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.093576 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/df650a21-47af-445a-9b8a-8908684aa6b9-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.093679 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df650a21-47af-445a-9b8a-8908684aa6b9-config\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.094307 4828 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.094350 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d81fdf23-e471-4a74-a4ac-24e665108494\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d81fdf23-e471-4a74-a4ac-24e665108494\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/56aaf1ead3b3052ce4968a9b50a2e9404bf7726dd15d27075c8361fe067578e9/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.094632 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3501ce8-e4b7-420d-b727-5ed3015f39d2-config\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.094984 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3501ce8-e4b7-420d-b727-5ed3015f39d2-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.095304 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/d3501ce8-e4b7-420d-b727-5ed3015f39d2-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.096108 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/d3501ce8-e4b7-420d-b727-5ed3015f39d2-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.098377 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/d3501ce8-e4b7-420d-b727-5ed3015f39d2-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.109978 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkqkn\" (UniqueName: \"kubernetes.io/projected/d3501ce8-e4b7-420d-b727-5ed3015f39d2-kube-api-access-mkqkn\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.117899 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d81fdf23-e471-4a74-a4ac-24e665108494\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d81fdf23-e471-4a74-a4ac-24e665108494\") pod \"logging-loki-compactor-0\" (UID: \"d3501ce8-e4b7-420d-b727-5ed3015f39d2\") " pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.122433 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.194599 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df650a21-47af-445a-9b8a-8908684aa6b9-config\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.194668 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df650a21-47af-445a-9b8a-8908684aa6b9-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.194748 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxzps\" (UniqueName: \"kubernetes.io/projected/df650a21-47af-445a-9b8a-8908684aa6b9-kube-api-access-hxzps\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.194781 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/df650a21-47af-445a-9b8a-8908684aa6b9-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.194818 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/df650a21-47af-445a-9b8a-8908684aa6b9-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.194889 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d3422407-00c4-4e72-96b2-cb91ea9c78e8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d3422407-00c4-4e72-96b2-cb91ea9c78e8\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.195356 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/df650a21-47af-445a-9b8a-8908684aa6b9-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.195677 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/df650a21-47af-445a-9b8a-8908684aa6b9-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.196104 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df650a21-47af-445a-9b8a-8908684aa6b9-config\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.204476 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/df650a21-47af-445a-9b8a-8908684aa6b9-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.214507 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/df650a21-47af-445a-9b8a-8908684aa6b9-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.214988 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/df650a21-47af-445a-9b8a-8908684aa6b9-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.215393 4828 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.215418 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d3422407-00c4-4e72-96b2-cb91ea9c78e8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d3422407-00c4-4e72-96b2-cb91ea9c78e8\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/376cb11a8cbbf967ead9ba58b25841021a00daa252f4ee0ccac3cbca7d885a3f/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.238117 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxzps\" (UniqueName: \"kubernetes.io/projected/df650a21-47af-445a-9b8a-8908684aa6b9-kube-api-access-hxzps\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.266097 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d3422407-00c4-4e72-96b2-cb91ea9c78e8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d3422407-00c4-4e72-96b2-cb91ea9c78e8\") pod \"logging-loki-index-gateway-0\" (UID: \"df650a21-47af-445a-9b8a-8908684aa6b9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.269265 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.283448 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.637813 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 10 19:10:02 crc kubenswrapper[4828]: W1210 19:10:02.643949 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4483694_bd44_4582_80b0_17c774d39682.slice/crio-9417ab73fa17336e57acd6ebd8f93eecfcff29ab18b2fa03cd0995d2483b89ab WatchSource:0}: Error finding container 9417ab73fa17336e57acd6ebd8f93eecfcff29ab18b2fa03cd0995d2483b89ab: Status 404 returned error can't find the container with id 9417ab73fa17336e57acd6ebd8f93eecfcff29ab18b2fa03cd0995d2483b89ab Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.723280 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 10 19:10:02 crc kubenswrapper[4828]: I1210 19:10:02.732094 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 10 19:10:03 crc kubenswrapper[4828]: I1210 19:10:03.002961 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"df650a21-47af-445a-9b8a-8908684aa6b9","Type":"ContainerStarted","Data":"4487ebaae3887587045c6d6d68e8d364e9dd9f89fd51536e9008f1b721be03a3"} Dec 10 19:10:03 crc kubenswrapper[4828]: I1210 19:10:03.004124 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"d3501ce8-e4b7-420d-b727-5ed3015f39d2","Type":"ContainerStarted","Data":"89f584a2ec39aab4d40fe0a4d4505d873cee81afb909008ec1d2b0880d75a440"} Dec 10 19:10:03 crc kubenswrapper[4828]: I1210 19:10:03.005337 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"a4483694-bd44-4582-80b0-17c774d39682","Type":"ContainerStarted","Data":"9417ab73fa17336e57acd6ebd8f93eecfcff29ab18b2fa03cd0995d2483b89ab"} Dec 10 19:10:07 crc kubenswrapper[4828]: I1210 19:10:07.049726 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" event={"ID":"14db470d-4202-4b2d-9339-65ba24e521d0","Type":"ContainerStarted","Data":"833c173bda4f1a3db49b6530efcb0590ef18a5f73085cf0cc9e6980a90f8bc1b"} Dec 10 19:10:07 crc kubenswrapper[4828]: I1210 19:10:07.050477 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:07 crc kubenswrapper[4828]: I1210 19:10:07.069280 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" event={"ID":"cf17217c-0dc6-4edd-8334-8896f103f9f1","Type":"ContainerStarted","Data":"7c09aac770f3c2cf20b458dddf08bf21a1855d0d48ebfe9b6caec06f21c93661"} Dec 10 19:10:08 crc kubenswrapper[4828]: I1210 19:10:08.078343 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"d3501ce8-e4b7-420d-b727-5ed3015f39d2","Type":"ContainerStarted","Data":"218b2d2824c58ce2d102967e74e0b06359128799f165e5de594a31c845a88419"} Dec 10 19:10:08 crc kubenswrapper[4828]: I1210 19:10:08.079955 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:08 crc kubenswrapper[4828]: I1210 19:10:08.080854 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"a4483694-bd44-4582-80b0-17c774d39682","Type":"ContainerStarted","Data":"3a61ceaf6106c13f4a16a32d2bd0ac8d46e771ccf6e695baa1ac82d8622802f5"} Dec 10 19:10:08 crc kubenswrapper[4828]: I1210 19:10:08.081060 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:10:08 crc kubenswrapper[4828]: I1210 19:10:08.083471 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" event={"ID":"6cf6784a-0699-4d1c-ba27-cb8f25ab729b","Type":"ContainerStarted","Data":"1aa069391467ff5aeca4cce4ebe2fdf6dfeb51861ac2fd1bb7043647596e1a4c"} Dec 10 19:10:08 crc kubenswrapper[4828]: I1210 19:10:08.083691 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:08 crc kubenswrapper[4828]: I1210 19:10:08.085633 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" event={"ID":"4c744ad4-db8c-42ad-9b49-ecb32b023cb6","Type":"ContainerStarted","Data":"7c372339d2d9627d8f0b70a286ad265d955d75f0ca2fadbb16750e963d955792"} Dec 10 19:10:08 crc kubenswrapper[4828]: I1210 19:10:08.087785 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"df650a21-47af-445a-9b8a-8908684aa6b9","Type":"ContainerStarted","Data":"cebec627b4afabb1fe1b43f23ccbdd6f96062f5386f3b4e311a471ded6f251f7"} Dec 10 19:10:08 crc kubenswrapper[4828]: I1210 19:10:08.087947 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:08 crc kubenswrapper[4828]: I1210 19:10:08.089792 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" event={"ID":"15695070-e7cd-4af1-980d-e79defa1efaf","Type":"ContainerStarted","Data":"d5f35bcfa3312a85a85e02e85faaf565587f3a4beaca409954a2c3e51e71c339"} Dec 10 19:10:08 crc kubenswrapper[4828]: I1210 19:10:08.090002 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:08 crc kubenswrapper[4828]: I1210 19:10:08.127980 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=4.048733413 podStartE2EDuration="8.127951116s" podCreationTimestamp="2025-12-10 19:10:00 +0000 UTC" firstStartedPulling="2025-12-10 19:10:02.729571509 +0000 UTC m=+883.240182544" lastFinishedPulling="2025-12-10 19:10:06.808789242 +0000 UTC m=+887.319400247" observedRunningTime="2025-12-10 19:10:08.123693726 +0000 UTC m=+888.634304741" watchObservedRunningTime="2025-12-10 19:10:08.127951116 +0000 UTC m=+888.638562161" Dec 10 19:10:08 crc kubenswrapper[4828]: I1210 19:10:08.132719 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" podStartSLOduration=2.847180747 podStartE2EDuration="8.13270124s" podCreationTimestamp="2025-12-10 19:10:00 +0000 UTC" firstStartedPulling="2025-12-10 19:10:01.462123134 +0000 UTC m=+881.972734149" lastFinishedPulling="2025-12-10 19:10:06.747643637 +0000 UTC m=+887.258254642" observedRunningTime="2025-12-10 19:10:07.074166131 +0000 UTC m=+887.584777136" watchObservedRunningTime="2025-12-10 19:10:08.13270124 +0000 UTC m=+888.643312285" Dec 10 19:10:08 crc kubenswrapper[4828]: I1210 19:10:08.148524 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" podStartSLOduration=2.8465987520000002 podStartE2EDuration="8.148506389s" podCreationTimestamp="2025-12-10 19:10:00 +0000 UTC" firstStartedPulling="2025-12-10 19:10:01.595868101 +0000 UTC m=+882.106479106" lastFinishedPulling="2025-12-10 19:10:06.897775748 +0000 UTC m=+887.408386743" observedRunningTime="2025-12-10 19:10:08.143999503 +0000 UTC m=+888.654610508" watchObservedRunningTime="2025-12-10 19:10:08.148506389 +0000 UTC m=+888.659117414" Dec 10 19:10:08 crc kubenswrapper[4828]: I1210 19:10:08.175000 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=4.013744446 podStartE2EDuration="8.174965135s" podCreationTimestamp="2025-12-10 19:10:00 +0000 UTC" firstStartedPulling="2025-12-10 19:10:02.645713725 +0000 UTC m=+883.156324730" lastFinishedPulling="2025-12-10 19:10:06.806934424 +0000 UTC m=+887.317545419" observedRunningTime="2025-12-10 19:10:08.170786357 +0000 UTC m=+888.681397372" watchObservedRunningTime="2025-12-10 19:10:08.174965135 +0000 UTC m=+888.685576180" Dec 10 19:10:08 crc kubenswrapper[4828]: I1210 19:10:08.193927 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" podStartSLOduration=3.072159008 podStartE2EDuration="8.193909206s" podCreationTimestamp="2025-12-10 19:10:00 +0000 UTC" firstStartedPulling="2025-12-10 19:10:01.686660874 +0000 UTC m=+882.197271879" lastFinishedPulling="2025-12-10 19:10:06.808411062 +0000 UTC m=+887.319022077" observedRunningTime="2025-12-10 19:10:08.186770281 +0000 UTC m=+888.697381286" watchObservedRunningTime="2025-12-10 19:10:08.193909206 +0000 UTC m=+888.704520211" Dec 10 19:10:08 crc kubenswrapper[4828]: I1210 19:10:08.210545 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=4.133250525 podStartE2EDuration="8.210522817s" podCreationTimestamp="2025-12-10 19:10:00 +0000 UTC" firstStartedPulling="2025-12-10 19:10:02.730140774 +0000 UTC m=+883.240751779" lastFinishedPulling="2025-12-10 19:10:06.807413066 +0000 UTC m=+887.318024071" observedRunningTime="2025-12-10 19:10:08.207525379 +0000 UTC m=+888.718136394" watchObservedRunningTime="2025-12-10 19:10:08.210522817 +0000 UTC m=+888.721133822" Dec 10 19:10:10 crc kubenswrapper[4828]: I1210 19:10:10.105266 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" event={"ID":"cf17217c-0dc6-4edd-8334-8896f103f9f1","Type":"ContainerStarted","Data":"73aa2c255d8f71c1d960c9352223d0a979767e00041f9bc297a597c92bafec04"} Dec 10 19:10:10 crc kubenswrapper[4828]: I1210 19:10:10.105895 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:10 crc kubenswrapper[4828]: I1210 19:10:10.107216 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" event={"ID":"4c744ad4-db8c-42ad-9b49-ecb32b023cb6","Type":"ContainerStarted","Data":"8d4f4f8a16d2e73d8aac112d4b836c8aad08184e5d620ccdc3516e7ddaf1ec09"} Dec 10 19:10:10 crc kubenswrapper[4828]: I1210 19:10:10.107526 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:10 crc kubenswrapper[4828]: I1210 19:10:10.107560 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:10 crc kubenswrapper[4828]: I1210 19:10:10.114480 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:10 crc kubenswrapper[4828]: I1210 19:10:10.114787 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:10 crc kubenswrapper[4828]: I1210 19:10:10.118599 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" Dec 10 19:10:10 crc kubenswrapper[4828]: I1210 19:10:10.129288 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" podStartSLOduration=2.754986708 podStartE2EDuration="10.129265495s" podCreationTimestamp="2025-12-10 19:10:00 +0000 UTC" firstStartedPulling="2025-12-10 19:10:01.917746525 +0000 UTC m=+882.428357530" lastFinishedPulling="2025-12-10 19:10:09.292025312 +0000 UTC m=+889.802636317" observedRunningTime="2025-12-10 19:10:10.12327897 +0000 UTC m=+890.633889975" watchObservedRunningTime="2025-12-10 19:10:10.129265495 +0000 UTC m=+890.639876500" Dec 10 19:10:10 crc kubenswrapper[4828]: I1210 19:10:10.191520 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-6c4d974455-7r2h2" podStartSLOduration=2.7007309619999997 podStartE2EDuration="10.191501729s" podCreationTimestamp="2025-12-10 19:10:00 +0000 UTC" firstStartedPulling="2025-12-10 19:10:01.851357314 +0000 UTC m=+882.361968319" lastFinishedPulling="2025-12-10 19:10:09.342128081 +0000 UTC m=+889.852739086" observedRunningTime="2025-12-10 19:10:10.189006404 +0000 UTC m=+890.699617409" watchObservedRunningTime="2025-12-10 19:10:10.191501729 +0000 UTC m=+890.702112734" Dec 10 19:10:11 crc kubenswrapper[4828]: I1210 19:10:11.114032 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:11 crc kubenswrapper[4828]: I1210 19:10:11.122688 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-6c4d974455-7lpjj" Dec 10 19:10:22 crc kubenswrapper[4828]: I1210 19:10:22.127469 4828 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Dec 10 19:10:22 crc kubenswrapper[4828]: I1210 19:10:22.128111 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="a4483694-bd44-4582-80b0-17c774d39682" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 10 19:10:22 crc kubenswrapper[4828]: I1210 19:10:22.275954 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Dec 10 19:10:22 crc kubenswrapper[4828]: I1210 19:10:22.292833 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Dec 10 19:10:30 crc kubenswrapper[4828]: I1210 19:10:30.961826 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-76cc67bf56-lg62n" Dec 10 19:10:31 crc kubenswrapper[4828]: I1210 19:10:31.131264 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-5895d59bb8-4pkdl" Dec 10 19:10:31 crc kubenswrapper[4828]: I1210 19:10:31.235022 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-qqdqb" Dec 10 19:10:32 crc kubenswrapper[4828]: I1210 19:10:32.127986 4828 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Dec 10 19:10:32 crc kubenswrapper[4828]: I1210 19:10:32.128355 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="a4483694-bd44-4582-80b0-17c774d39682" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 10 19:10:32 crc kubenswrapper[4828]: I1210 19:10:32.780466 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4rhgw"] Dec 10 19:10:32 crc kubenswrapper[4828]: I1210 19:10:32.782674 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4rhgw" Dec 10 19:10:32 crc kubenswrapper[4828]: I1210 19:10:32.798693 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4rhgw"] Dec 10 19:10:32 crc kubenswrapper[4828]: I1210 19:10:32.885624 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52h72\" (UniqueName: \"kubernetes.io/projected/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-kube-api-access-52h72\") pod \"certified-operators-4rhgw\" (UID: \"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065\") " pod="openshift-marketplace/certified-operators-4rhgw" Dec 10 19:10:32 crc kubenswrapper[4828]: I1210 19:10:32.885735 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-catalog-content\") pod \"certified-operators-4rhgw\" (UID: \"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065\") " pod="openshift-marketplace/certified-operators-4rhgw" Dec 10 19:10:32 crc kubenswrapper[4828]: I1210 19:10:32.885940 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-utilities\") pod \"certified-operators-4rhgw\" (UID: \"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065\") " pod="openshift-marketplace/certified-operators-4rhgw" Dec 10 19:10:32 crc kubenswrapper[4828]: I1210 19:10:32.987672 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-catalog-content\") pod \"certified-operators-4rhgw\" (UID: \"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065\") " pod="openshift-marketplace/certified-operators-4rhgw" Dec 10 19:10:32 crc kubenswrapper[4828]: I1210 19:10:32.987792 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-utilities\") pod \"certified-operators-4rhgw\" (UID: \"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065\") " pod="openshift-marketplace/certified-operators-4rhgw" Dec 10 19:10:32 crc kubenswrapper[4828]: I1210 19:10:32.987844 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52h72\" (UniqueName: \"kubernetes.io/projected/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-kube-api-access-52h72\") pod \"certified-operators-4rhgw\" (UID: \"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065\") " pod="openshift-marketplace/certified-operators-4rhgw" Dec 10 19:10:32 crc kubenswrapper[4828]: I1210 19:10:32.988439 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-utilities\") pod \"certified-operators-4rhgw\" (UID: \"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065\") " pod="openshift-marketplace/certified-operators-4rhgw" Dec 10 19:10:32 crc kubenswrapper[4828]: I1210 19:10:32.988482 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-catalog-content\") pod \"certified-operators-4rhgw\" (UID: \"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065\") " pod="openshift-marketplace/certified-operators-4rhgw" Dec 10 19:10:33 crc kubenswrapper[4828]: I1210 19:10:33.009087 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52h72\" (UniqueName: \"kubernetes.io/projected/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-kube-api-access-52h72\") pod \"certified-operators-4rhgw\" (UID: \"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065\") " pod="openshift-marketplace/certified-operators-4rhgw" Dec 10 19:10:33 crc kubenswrapper[4828]: I1210 19:10:33.105392 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4rhgw" Dec 10 19:10:33 crc kubenswrapper[4828]: I1210 19:10:33.655001 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4rhgw"] Dec 10 19:10:34 crc kubenswrapper[4828]: I1210 19:10:34.294099 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rhgw" event={"ID":"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065","Type":"ContainerStarted","Data":"a888d6a7c202016dfb37e12097e7b305dd3f5079c0ebe0ac50f4a60bc6b69e77"} Dec 10 19:10:35 crc kubenswrapper[4828]: I1210 19:10:35.301557 4828 generic.go:334] "Generic (PLEG): container finished" podID="5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065" containerID="4d409d3cf1a68c6992d396dae4551433001fea8594dbf317e4f476da2b58b141" exitCode=0 Dec 10 19:10:35 crc kubenswrapper[4828]: I1210 19:10:35.301607 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rhgw" event={"ID":"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065","Type":"ContainerDied","Data":"4d409d3cf1a68c6992d396dae4551433001fea8594dbf317e4f476da2b58b141"} Dec 10 19:10:42 crc kubenswrapper[4828]: I1210 19:10:42.126938 4828 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 10 19:10:42 crc kubenswrapper[4828]: I1210 19:10:42.127609 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="a4483694-bd44-4582-80b0-17c774d39682" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 10 19:10:44 crc kubenswrapper[4828]: I1210 19:10:44.361174 4828 generic.go:334] "Generic (PLEG): container finished" podID="5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065" containerID="e1de59dac8272736eefcba4bf88453d4a852b5d12c59810902a1373b38571b69" exitCode=0 Dec 10 19:10:44 crc kubenswrapper[4828]: I1210 19:10:44.361378 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rhgw" event={"ID":"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065","Type":"ContainerDied","Data":"e1de59dac8272736eefcba4bf88453d4a852b5d12c59810902a1373b38571b69"} Dec 10 19:10:45 crc kubenswrapper[4828]: I1210 19:10:45.079209 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m7bgg"] Dec 10 19:10:45 crc kubenswrapper[4828]: I1210 19:10:45.080898 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m7bgg" Dec 10 19:10:45 crc kubenswrapper[4828]: I1210 19:10:45.178891 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c52880d-4bb5-4665-964c-b378edcb5ac9-catalog-content\") pod \"redhat-marketplace-m7bgg\" (UID: \"7c52880d-4bb5-4665-964c-b378edcb5ac9\") " pod="openshift-marketplace/redhat-marketplace-m7bgg" Dec 10 19:10:45 crc kubenswrapper[4828]: I1210 19:10:45.178972 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rghjz\" (UniqueName: \"kubernetes.io/projected/7c52880d-4bb5-4665-964c-b378edcb5ac9-kube-api-access-rghjz\") pod \"redhat-marketplace-m7bgg\" (UID: \"7c52880d-4bb5-4665-964c-b378edcb5ac9\") " pod="openshift-marketplace/redhat-marketplace-m7bgg" Dec 10 19:10:45 crc kubenswrapper[4828]: I1210 19:10:45.179223 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c52880d-4bb5-4665-964c-b378edcb5ac9-utilities\") pod \"redhat-marketplace-m7bgg\" (UID: \"7c52880d-4bb5-4665-964c-b378edcb5ac9\") " pod="openshift-marketplace/redhat-marketplace-m7bgg" Dec 10 19:10:45 crc kubenswrapper[4828]: I1210 19:10:45.280323 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rghjz\" (UniqueName: \"kubernetes.io/projected/7c52880d-4bb5-4665-964c-b378edcb5ac9-kube-api-access-rghjz\") pod \"redhat-marketplace-m7bgg\" (UID: \"7c52880d-4bb5-4665-964c-b378edcb5ac9\") " pod="openshift-marketplace/redhat-marketplace-m7bgg" Dec 10 19:10:45 crc kubenswrapper[4828]: I1210 19:10:45.280480 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c52880d-4bb5-4665-964c-b378edcb5ac9-utilities\") pod \"redhat-marketplace-m7bgg\" (UID: \"7c52880d-4bb5-4665-964c-b378edcb5ac9\") " pod="openshift-marketplace/redhat-marketplace-m7bgg" Dec 10 19:10:45 crc kubenswrapper[4828]: I1210 19:10:45.280552 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c52880d-4bb5-4665-964c-b378edcb5ac9-catalog-content\") pod \"redhat-marketplace-m7bgg\" (UID: \"7c52880d-4bb5-4665-964c-b378edcb5ac9\") " pod="openshift-marketplace/redhat-marketplace-m7bgg" Dec 10 19:10:45 crc kubenswrapper[4828]: I1210 19:10:45.281019 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c52880d-4bb5-4665-964c-b378edcb5ac9-utilities\") pod \"redhat-marketplace-m7bgg\" (UID: \"7c52880d-4bb5-4665-964c-b378edcb5ac9\") " pod="openshift-marketplace/redhat-marketplace-m7bgg" Dec 10 19:10:45 crc kubenswrapper[4828]: I1210 19:10:45.281115 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c52880d-4bb5-4665-964c-b378edcb5ac9-catalog-content\") pod \"redhat-marketplace-m7bgg\" (UID: \"7c52880d-4bb5-4665-964c-b378edcb5ac9\") " pod="openshift-marketplace/redhat-marketplace-m7bgg" Dec 10 19:10:45 crc kubenswrapper[4828]: I1210 19:10:45.307520 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rghjz\" (UniqueName: \"kubernetes.io/projected/7c52880d-4bb5-4665-964c-b378edcb5ac9-kube-api-access-rghjz\") pod \"redhat-marketplace-m7bgg\" (UID: \"7c52880d-4bb5-4665-964c-b378edcb5ac9\") " pod="openshift-marketplace/redhat-marketplace-m7bgg" Dec 10 19:10:45 crc kubenswrapper[4828]: I1210 19:10:45.397164 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m7bgg" Dec 10 19:10:45 crc kubenswrapper[4828]: I1210 19:10:45.594417 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m7bgg"] Dec 10 19:10:46 crc kubenswrapper[4828]: I1210 19:10:46.064576 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m7bgg"] Dec 10 19:10:46 crc kubenswrapper[4828]: W1210 19:10:46.067824 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c52880d_4bb5_4665_964c_b378edcb5ac9.slice/crio-e64360578f885d5789d941b870db8a313fa0a6acd5f578fbcb119dab9771cc06 WatchSource:0}: Error finding container e64360578f885d5789d941b870db8a313fa0a6acd5f578fbcb119dab9771cc06: Status 404 returned error can't find the container with id e64360578f885d5789d941b870db8a313fa0a6acd5f578fbcb119dab9771cc06 Dec 10 19:10:46 crc kubenswrapper[4828]: I1210 19:10:46.374007 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7bgg" event={"ID":"7c52880d-4bb5-4665-964c-b378edcb5ac9","Type":"ContainerStarted","Data":"e64360578f885d5789d941b870db8a313fa0a6acd5f578fbcb119dab9771cc06"} Dec 10 19:10:47 crc kubenswrapper[4828]: I1210 19:10:47.381375 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7bgg" event={"ID":"7c52880d-4bb5-4665-964c-b378edcb5ac9","Type":"ContainerStarted","Data":"db57975b70b2b46093a4a150e4c83220dbdd5690ae0b4eda9ed5b777d57ad535"} Dec 10 19:10:48 crc kubenswrapper[4828]: I1210 19:10:48.389770 4828 generic.go:334] "Generic (PLEG): container finished" podID="7c52880d-4bb5-4665-964c-b378edcb5ac9" containerID="db57975b70b2b46093a4a150e4c83220dbdd5690ae0b4eda9ed5b777d57ad535" exitCode=0 Dec 10 19:10:48 crc kubenswrapper[4828]: I1210 19:10:48.389840 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7bgg" event={"ID":"7c52880d-4bb5-4665-964c-b378edcb5ac9","Type":"ContainerDied","Data":"db57975b70b2b46093a4a150e4c83220dbdd5690ae0b4eda9ed5b777d57ad535"} Dec 10 19:10:48 crc kubenswrapper[4828]: I1210 19:10:48.393336 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rhgw" event={"ID":"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065","Type":"ContainerStarted","Data":"e17c44148dab8622b13de1c71a7e901317109e245eaecad0fd3b4301c53c28c1"} Dec 10 19:10:48 crc kubenswrapper[4828]: I1210 19:10:48.431477 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4rhgw" podStartSLOduration=3.8063928479999998 podStartE2EDuration="16.431456402s" podCreationTimestamp="2025-12-10 19:10:32 +0000 UTC" firstStartedPulling="2025-12-10 19:10:35.303382764 +0000 UTC m=+915.813993769" lastFinishedPulling="2025-12-10 19:10:47.928446298 +0000 UTC m=+928.439057323" observedRunningTime="2025-12-10 19:10:48.425639701 +0000 UTC m=+928.936250706" watchObservedRunningTime="2025-12-10 19:10:48.431456402 +0000 UTC m=+928.942067417" Dec 10 19:10:51 crc kubenswrapper[4828]: I1210 19:10:51.231087 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:10:51 crc kubenswrapper[4828]: I1210 19:10:51.231458 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:10:52 crc kubenswrapper[4828]: I1210 19:10:52.130333 4828 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 10 19:10:52 crc kubenswrapper[4828]: I1210 19:10:52.130427 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="a4483694-bd44-4582-80b0-17c774d39682" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 10 19:10:53 crc kubenswrapper[4828]: I1210 19:10:53.106214 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4rhgw" Dec 10 19:10:53 crc kubenswrapper[4828]: I1210 19:10:53.106554 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4rhgw" Dec 10 19:10:53 crc kubenswrapper[4828]: I1210 19:10:53.169003 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4rhgw" Dec 10 19:10:53 crc kubenswrapper[4828]: I1210 19:10:53.493081 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4rhgw" Dec 10 19:10:53 crc kubenswrapper[4828]: I1210 19:10:53.553571 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4rhgw"] Dec 10 19:10:55 crc kubenswrapper[4828]: I1210 19:10:55.447441 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4rhgw" podUID="5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065" containerName="registry-server" containerID="cri-o://e17c44148dab8622b13de1c71a7e901317109e245eaecad0fd3b4301c53c28c1" gracePeriod=2 Dec 10 19:10:57 crc kubenswrapper[4828]: I1210 19:10:57.465096 4828 generic.go:334] "Generic (PLEG): container finished" podID="5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065" containerID="e17c44148dab8622b13de1c71a7e901317109e245eaecad0fd3b4301c53c28c1" exitCode=0 Dec 10 19:10:57 crc kubenswrapper[4828]: I1210 19:10:57.465174 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rhgw" event={"ID":"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065","Type":"ContainerDied","Data":"e17c44148dab8622b13de1c71a7e901317109e245eaecad0fd3b4301c53c28c1"} Dec 10 19:10:57 crc kubenswrapper[4828]: I1210 19:10:57.465615 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4rhgw" event={"ID":"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065","Type":"ContainerDied","Data":"a888d6a7c202016dfb37e12097e7b305dd3f5079c0ebe0ac50f4a60bc6b69e77"} Dec 10 19:10:57 crc kubenswrapper[4828]: I1210 19:10:57.465631 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a888d6a7c202016dfb37e12097e7b305dd3f5079c0ebe0ac50f4a60bc6b69e77" Dec 10 19:10:57 crc kubenswrapper[4828]: I1210 19:10:57.467553 4828 generic.go:334] "Generic (PLEG): container finished" podID="7c52880d-4bb5-4665-964c-b378edcb5ac9" containerID="9d2de2fb49441d9b477867b9a7c644a2855c875966dc78fb896d2ef570ecfdef" exitCode=0 Dec 10 19:10:57 crc kubenswrapper[4828]: I1210 19:10:57.467576 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7bgg" event={"ID":"7c52880d-4bb5-4665-964c-b378edcb5ac9","Type":"ContainerDied","Data":"9d2de2fb49441d9b477867b9a7c644a2855c875966dc78fb896d2ef570ecfdef"} Dec 10 19:10:57 crc kubenswrapper[4828]: I1210 19:10:57.499680 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4rhgw" Dec 10 19:10:57 crc kubenswrapper[4828]: I1210 19:10:57.580868 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-catalog-content\") pod \"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065\" (UID: \"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065\") " Dec 10 19:10:57 crc kubenswrapper[4828]: I1210 19:10:57.581003 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52h72\" (UniqueName: \"kubernetes.io/projected/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-kube-api-access-52h72\") pod \"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065\" (UID: \"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065\") " Dec 10 19:10:57 crc kubenswrapper[4828]: I1210 19:10:57.581076 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-utilities\") pod \"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065\" (UID: \"5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065\") " Dec 10 19:10:57 crc kubenswrapper[4828]: I1210 19:10:57.582526 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-utilities" (OuterVolumeSpecName: "utilities") pod "5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065" (UID: "5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:10:57 crc kubenswrapper[4828]: I1210 19:10:57.587187 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-kube-api-access-52h72" (OuterVolumeSpecName: "kube-api-access-52h72") pod "5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065" (UID: "5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065"). InnerVolumeSpecName "kube-api-access-52h72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:10:57 crc kubenswrapper[4828]: I1210 19:10:57.629118 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065" (UID: "5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:10:57 crc kubenswrapper[4828]: I1210 19:10:57.683068 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:57 crc kubenswrapper[4828]: I1210 19:10:57.683102 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:57 crc kubenswrapper[4828]: I1210 19:10:57.683116 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52h72\" (UniqueName: \"kubernetes.io/projected/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065-kube-api-access-52h72\") on node \"crc\" DevicePath \"\"" Dec 10 19:10:58 crc kubenswrapper[4828]: I1210 19:10:58.473238 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4rhgw" Dec 10 19:10:58 crc kubenswrapper[4828]: I1210 19:10:58.492550 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4rhgw"] Dec 10 19:10:58 crc kubenswrapper[4828]: I1210 19:10:58.499344 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4rhgw"] Dec 10 19:10:59 crc kubenswrapper[4828]: I1210 19:10:59.797326 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065" path="/var/lib/kubelet/pods/5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065/volumes" Dec 10 19:11:02 crc kubenswrapper[4828]: I1210 19:11:02.128909 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Dec 10 19:11:05 crc kubenswrapper[4828]: I1210 19:11:05.525838 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7bgg" event={"ID":"7c52880d-4bb5-4665-964c-b378edcb5ac9","Type":"ContainerStarted","Data":"692b62ce21b1b035fb19b964a5b4025b180d3cea2b419c79f2d05e3c4b242a53"} Dec 10 19:11:05 crc kubenswrapper[4828]: I1210 19:11:05.545137 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m7bgg" podStartSLOduration=10.704155725 podStartE2EDuration="20.545121001s" podCreationTimestamp="2025-12-10 19:10:45 +0000 UTC" firstStartedPulling="2025-12-10 19:10:48.392304344 +0000 UTC m=+928.902915349" lastFinishedPulling="2025-12-10 19:10:58.2332696 +0000 UTC m=+938.743880625" observedRunningTime="2025-12-10 19:11:05.540237814 +0000 UTC m=+946.050848839" watchObservedRunningTime="2025-12-10 19:11:05.545121001 +0000 UTC m=+946.055732006" Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.111120 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-47724"] Dec 10 19:11:06 crc kubenswrapper[4828]: E1210 19:11:06.111447 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065" containerName="registry-server" Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.111467 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065" containerName="registry-server" Dec 10 19:11:06 crc kubenswrapper[4828]: E1210 19:11:06.111491 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065" containerName="extract-utilities" Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.111500 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065" containerName="extract-utilities" Dec 10 19:11:06 crc kubenswrapper[4828]: E1210 19:11:06.111528 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065" containerName="extract-content" Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.111536 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065" containerName="extract-content" Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.111695 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ad3e6b2-d3ea-4ad5-a8fb-795c0f4d9065" containerName="registry-server" Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.112910 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-47724" Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.122048 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-47724"] Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.222973 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nf7r\" (UniqueName: \"kubernetes.io/projected/d3391181-ec59-42b3-b7a7-8f023bb54b9d-kube-api-access-5nf7r\") pod \"community-operators-47724\" (UID: \"d3391181-ec59-42b3-b7a7-8f023bb54b9d\") " pod="openshift-marketplace/community-operators-47724" Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.223054 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3391181-ec59-42b3-b7a7-8f023bb54b9d-utilities\") pod \"community-operators-47724\" (UID: \"d3391181-ec59-42b3-b7a7-8f023bb54b9d\") " pod="openshift-marketplace/community-operators-47724" Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.223096 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3391181-ec59-42b3-b7a7-8f023bb54b9d-catalog-content\") pod \"community-operators-47724\" (UID: \"d3391181-ec59-42b3-b7a7-8f023bb54b9d\") " pod="openshift-marketplace/community-operators-47724" Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.324458 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nf7r\" (UniqueName: \"kubernetes.io/projected/d3391181-ec59-42b3-b7a7-8f023bb54b9d-kube-api-access-5nf7r\") pod \"community-operators-47724\" (UID: \"d3391181-ec59-42b3-b7a7-8f023bb54b9d\") " pod="openshift-marketplace/community-operators-47724" Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.324576 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3391181-ec59-42b3-b7a7-8f023bb54b9d-utilities\") pod \"community-operators-47724\" (UID: \"d3391181-ec59-42b3-b7a7-8f023bb54b9d\") " pod="openshift-marketplace/community-operators-47724" Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.324621 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3391181-ec59-42b3-b7a7-8f023bb54b9d-catalog-content\") pod \"community-operators-47724\" (UID: \"d3391181-ec59-42b3-b7a7-8f023bb54b9d\") " pod="openshift-marketplace/community-operators-47724" Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.325145 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3391181-ec59-42b3-b7a7-8f023bb54b9d-utilities\") pod \"community-operators-47724\" (UID: \"d3391181-ec59-42b3-b7a7-8f023bb54b9d\") " pod="openshift-marketplace/community-operators-47724" Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.325258 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3391181-ec59-42b3-b7a7-8f023bb54b9d-catalog-content\") pod \"community-operators-47724\" (UID: \"d3391181-ec59-42b3-b7a7-8f023bb54b9d\") " pod="openshift-marketplace/community-operators-47724" Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.346092 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nf7r\" (UniqueName: \"kubernetes.io/projected/d3391181-ec59-42b3-b7a7-8f023bb54b9d-kube-api-access-5nf7r\") pod \"community-operators-47724\" (UID: \"d3391181-ec59-42b3-b7a7-8f023bb54b9d\") " pod="openshift-marketplace/community-operators-47724" Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.431134 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-47724" Dec 10 19:11:06 crc kubenswrapper[4828]: I1210 19:11:06.878237 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-47724"] Dec 10 19:11:07 crc kubenswrapper[4828]: I1210 19:11:07.542527 4828 generic.go:334] "Generic (PLEG): container finished" podID="d3391181-ec59-42b3-b7a7-8f023bb54b9d" containerID="1d5aa02cd227e6ed8c626647ccdd2216d7675977d771c05587558c0576db0fda" exitCode=0 Dec 10 19:11:07 crc kubenswrapper[4828]: I1210 19:11:07.542622 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-47724" event={"ID":"d3391181-ec59-42b3-b7a7-8f023bb54b9d","Type":"ContainerDied","Data":"1d5aa02cd227e6ed8c626647ccdd2216d7675977d771c05587558c0576db0fda"} Dec 10 19:11:07 crc kubenswrapper[4828]: I1210 19:11:07.542900 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-47724" event={"ID":"d3391181-ec59-42b3-b7a7-8f023bb54b9d","Type":"ContainerStarted","Data":"6ac36942102c5855fe842e8202a061a77829cfcdafd4c94c983285f7de6269d5"} Dec 10 19:11:09 crc kubenswrapper[4828]: I1210 19:11:09.556791 4828 generic.go:334] "Generic (PLEG): container finished" podID="d3391181-ec59-42b3-b7a7-8f023bb54b9d" containerID="85e65ab5d398bc47589152b08a24487c16581bec769401e1cb7267be1967acc5" exitCode=0 Dec 10 19:11:09 crc kubenswrapper[4828]: I1210 19:11:09.556851 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-47724" event={"ID":"d3391181-ec59-42b3-b7a7-8f023bb54b9d","Type":"ContainerDied","Data":"85e65ab5d398bc47589152b08a24487c16581bec769401e1cb7267be1967acc5"} Dec 10 19:11:10 crc kubenswrapper[4828]: I1210 19:11:10.567721 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-47724" event={"ID":"d3391181-ec59-42b3-b7a7-8f023bb54b9d","Type":"ContainerStarted","Data":"c13d3fae90e15b4a5893345738a6f3493a848ed58be7d956a35ffdd912ded384"} Dec 10 19:11:10 crc kubenswrapper[4828]: I1210 19:11:10.584211 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-47724" podStartSLOduration=1.988380754 podStartE2EDuration="4.584194024s" podCreationTimestamp="2025-12-10 19:11:06 +0000 UTC" firstStartedPulling="2025-12-10 19:11:07.544024035 +0000 UTC m=+948.054635040" lastFinishedPulling="2025-12-10 19:11:10.139837305 +0000 UTC m=+950.650448310" observedRunningTime="2025-12-10 19:11:10.583106945 +0000 UTC m=+951.093717980" watchObservedRunningTime="2025-12-10 19:11:10.584194024 +0000 UTC m=+951.094805039" Dec 10 19:11:15 crc kubenswrapper[4828]: I1210 19:11:15.397822 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m7bgg" Dec 10 19:11:15 crc kubenswrapper[4828]: I1210 19:11:15.398155 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m7bgg" Dec 10 19:11:15 crc kubenswrapper[4828]: I1210 19:11:15.448562 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m7bgg" Dec 10 19:11:15 crc kubenswrapper[4828]: I1210 19:11:15.667582 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m7bgg" Dec 10 19:11:16 crc kubenswrapper[4828]: I1210 19:11:16.282412 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m7bgg"] Dec 10 19:11:16 crc kubenswrapper[4828]: I1210 19:11:16.431581 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-47724" Dec 10 19:11:16 crc kubenswrapper[4828]: I1210 19:11:16.431668 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-47724" Dec 10 19:11:16 crc kubenswrapper[4828]: I1210 19:11:16.485298 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-47724" Dec 10 19:11:16 crc kubenswrapper[4828]: I1210 19:11:16.673107 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-47724" Dec 10 19:11:17 crc kubenswrapper[4828]: I1210 19:11:17.627293 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-m7bgg" podUID="7c52880d-4bb5-4665-964c-b378edcb5ac9" containerName="registry-server" containerID="cri-o://692b62ce21b1b035fb19b964a5b4025b180d3cea2b419c79f2d05e3c4b242a53" gracePeriod=2 Dec 10 19:11:18 crc kubenswrapper[4828]: I1210 19:11:18.634761 4828 generic.go:334] "Generic (PLEG): container finished" podID="7c52880d-4bb5-4665-964c-b378edcb5ac9" containerID="692b62ce21b1b035fb19b964a5b4025b180d3cea2b419c79f2d05e3c4b242a53" exitCode=0 Dec 10 19:11:18 crc kubenswrapper[4828]: I1210 19:11:18.634857 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7bgg" event={"ID":"7c52880d-4bb5-4665-964c-b378edcb5ac9","Type":"ContainerDied","Data":"692b62ce21b1b035fb19b964a5b4025b180d3cea2b419c79f2d05e3c4b242a53"} Dec 10 19:11:18 crc kubenswrapper[4828]: I1210 19:11:18.681943 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-47724"] Dec 10 19:11:18 crc kubenswrapper[4828]: I1210 19:11:18.682147 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-47724" podUID="d3391181-ec59-42b3-b7a7-8f023bb54b9d" containerName="registry-server" containerID="cri-o://c13d3fae90e15b4a5893345738a6f3493a848ed58be7d956a35ffdd912ded384" gracePeriod=2 Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.177491 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-47724" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.182824 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m7bgg" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.320372 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c52880d-4bb5-4665-964c-b378edcb5ac9-catalog-content\") pod \"7c52880d-4bb5-4665-964c-b378edcb5ac9\" (UID: \"7c52880d-4bb5-4665-964c-b378edcb5ac9\") " Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.320689 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3391181-ec59-42b3-b7a7-8f023bb54b9d-catalog-content\") pod \"d3391181-ec59-42b3-b7a7-8f023bb54b9d\" (UID: \"d3391181-ec59-42b3-b7a7-8f023bb54b9d\") " Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.320818 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rghjz\" (UniqueName: \"kubernetes.io/projected/7c52880d-4bb5-4665-964c-b378edcb5ac9-kube-api-access-rghjz\") pod \"7c52880d-4bb5-4665-964c-b378edcb5ac9\" (UID: \"7c52880d-4bb5-4665-964c-b378edcb5ac9\") " Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.320908 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c52880d-4bb5-4665-964c-b378edcb5ac9-utilities\") pod \"7c52880d-4bb5-4665-964c-b378edcb5ac9\" (UID: \"7c52880d-4bb5-4665-964c-b378edcb5ac9\") " Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.321057 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3391181-ec59-42b3-b7a7-8f023bb54b9d-utilities\") pod \"d3391181-ec59-42b3-b7a7-8f023bb54b9d\" (UID: \"d3391181-ec59-42b3-b7a7-8f023bb54b9d\") " Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.321166 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nf7r\" (UniqueName: \"kubernetes.io/projected/d3391181-ec59-42b3-b7a7-8f023bb54b9d-kube-api-access-5nf7r\") pod \"d3391181-ec59-42b3-b7a7-8f023bb54b9d\" (UID: \"d3391181-ec59-42b3-b7a7-8f023bb54b9d\") " Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.321667 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c52880d-4bb5-4665-964c-b378edcb5ac9-utilities" (OuterVolumeSpecName: "utilities") pod "7c52880d-4bb5-4665-964c-b378edcb5ac9" (UID: "7c52880d-4bb5-4665-964c-b378edcb5ac9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.321905 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3391181-ec59-42b3-b7a7-8f023bb54b9d-utilities" (OuterVolumeSpecName: "utilities") pod "d3391181-ec59-42b3-b7a7-8f023bb54b9d" (UID: "d3391181-ec59-42b3-b7a7-8f023bb54b9d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.325953 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c52880d-4bb5-4665-964c-b378edcb5ac9-kube-api-access-rghjz" (OuterVolumeSpecName: "kube-api-access-rghjz") pod "7c52880d-4bb5-4665-964c-b378edcb5ac9" (UID: "7c52880d-4bb5-4665-964c-b378edcb5ac9"). InnerVolumeSpecName "kube-api-access-rghjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.326458 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3391181-ec59-42b3-b7a7-8f023bb54b9d-kube-api-access-5nf7r" (OuterVolumeSpecName: "kube-api-access-5nf7r") pod "d3391181-ec59-42b3-b7a7-8f023bb54b9d" (UID: "d3391181-ec59-42b3-b7a7-8f023bb54b9d"). InnerVolumeSpecName "kube-api-access-5nf7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.343626 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c52880d-4bb5-4665-964c-b378edcb5ac9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c52880d-4bb5-4665-964c-b378edcb5ac9" (UID: "7c52880d-4bb5-4665-964c-b378edcb5ac9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.374478 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3391181-ec59-42b3-b7a7-8f023bb54b9d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3391181-ec59-42b3-b7a7-8f023bb54b9d" (UID: "d3391181-ec59-42b3-b7a7-8f023bb54b9d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.422514 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3391181-ec59-42b3-b7a7-8f023bb54b9d-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.422562 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nf7r\" (UniqueName: \"kubernetes.io/projected/d3391181-ec59-42b3-b7a7-8f023bb54b9d-kube-api-access-5nf7r\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.422574 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c52880d-4bb5-4665-964c-b378edcb5ac9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.422584 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3391181-ec59-42b3-b7a7-8f023bb54b9d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.422596 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rghjz\" (UniqueName: \"kubernetes.io/projected/7c52880d-4bb5-4665-964c-b378edcb5ac9-kube-api-access-rghjz\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.422608 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c52880d-4bb5-4665-964c-b378edcb5ac9-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.645109 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m7bgg" event={"ID":"7c52880d-4bb5-4665-964c-b378edcb5ac9","Type":"ContainerDied","Data":"e64360578f885d5789d941b870db8a313fa0a6acd5f578fbcb119dab9771cc06"} Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.646383 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m7bgg" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.646437 4828 scope.go:117] "RemoveContainer" containerID="692b62ce21b1b035fb19b964a5b4025b180d3cea2b419c79f2d05e3c4b242a53" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.656923 4828 generic.go:334] "Generic (PLEG): container finished" podID="d3391181-ec59-42b3-b7a7-8f023bb54b9d" containerID="c13d3fae90e15b4a5893345738a6f3493a848ed58be7d956a35ffdd912ded384" exitCode=0 Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.656973 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-47724" event={"ID":"d3391181-ec59-42b3-b7a7-8f023bb54b9d","Type":"ContainerDied","Data":"c13d3fae90e15b4a5893345738a6f3493a848ed58be7d956a35ffdd912ded384"} Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.657019 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-47724" event={"ID":"d3391181-ec59-42b3-b7a7-8f023bb54b9d","Type":"ContainerDied","Data":"6ac36942102c5855fe842e8202a061a77829cfcdafd4c94c983285f7de6269d5"} Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.657106 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-47724" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.667749 4828 scope.go:117] "RemoveContainer" containerID="9d2de2fb49441d9b477867b9a7c644a2855c875966dc78fb896d2ef570ecfdef" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.679071 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m7bgg"] Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.686565 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-m7bgg"] Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.695746 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-47724"] Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.700366 4828 scope.go:117] "RemoveContainer" containerID="db57975b70b2b46093a4a150e4c83220dbdd5690ae0b4eda9ed5b777d57ad535" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.704606 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-47724"] Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.716655 4828 scope.go:117] "RemoveContainer" containerID="c13d3fae90e15b4a5893345738a6f3493a848ed58be7d956a35ffdd912ded384" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.730358 4828 scope.go:117] "RemoveContainer" containerID="85e65ab5d398bc47589152b08a24487c16581bec769401e1cb7267be1967acc5" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.747324 4828 scope.go:117] "RemoveContainer" containerID="1d5aa02cd227e6ed8c626647ccdd2216d7675977d771c05587558c0576db0fda" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.761409 4828 scope.go:117] "RemoveContainer" containerID="c13d3fae90e15b4a5893345738a6f3493a848ed58be7d956a35ffdd912ded384" Dec 10 19:11:19 crc kubenswrapper[4828]: E1210 19:11:19.761828 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c13d3fae90e15b4a5893345738a6f3493a848ed58be7d956a35ffdd912ded384\": container with ID starting with c13d3fae90e15b4a5893345738a6f3493a848ed58be7d956a35ffdd912ded384 not found: ID does not exist" containerID="c13d3fae90e15b4a5893345738a6f3493a848ed58be7d956a35ffdd912ded384" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.761880 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c13d3fae90e15b4a5893345738a6f3493a848ed58be7d956a35ffdd912ded384"} err="failed to get container status \"c13d3fae90e15b4a5893345738a6f3493a848ed58be7d956a35ffdd912ded384\": rpc error: code = NotFound desc = could not find container \"c13d3fae90e15b4a5893345738a6f3493a848ed58be7d956a35ffdd912ded384\": container with ID starting with c13d3fae90e15b4a5893345738a6f3493a848ed58be7d956a35ffdd912ded384 not found: ID does not exist" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.761912 4828 scope.go:117] "RemoveContainer" containerID="85e65ab5d398bc47589152b08a24487c16581bec769401e1cb7267be1967acc5" Dec 10 19:11:19 crc kubenswrapper[4828]: E1210 19:11:19.762233 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85e65ab5d398bc47589152b08a24487c16581bec769401e1cb7267be1967acc5\": container with ID starting with 85e65ab5d398bc47589152b08a24487c16581bec769401e1cb7267be1967acc5 not found: ID does not exist" containerID="85e65ab5d398bc47589152b08a24487c16581bec769401e1cb7267be1967acc5" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.762272 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85e65ab5d398bc47589152b08a24487c16581bec769401e1cb7267be1967acc5"} err="failed to get container status \"85e65ab5d398bc47589152b08a24487c16581bec769401e1cb7267be1967acc5\": rpc error: code = NotFound desc = could not find container \"85e65ab5d398bc47589152b08a24487c16581bec769401e1cb7267be1967acc5\": container with ID starting with 85e65ab5d398bc47589152b08a24487c16581bec769401e1cb7267be1967acc5 not found: ID does not exist" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.762298 4828 scope.go:117] "RemoveContainer" containerID="1d5aa02cd227e6ed8c626647ccdd2216d7675977d771c05587558c0576db0fda" Dec 10 19:11:19 crc kubenswrapper[4828]: E1210 19:11:19.762721 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d5aa02cd227e6ed8c626647ccdd2216d7675977d771c05587558c0576db0fda\": container with ID starting with 1d5aa02cd227e6ed8c626647ccdd2216d7675977d771c05587558c0576db0fda not found: ID does not exist" containerID="1d5aa02cd227e6ed8c626647ccdd2216d7675977d771c05587558c0576db0fda" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.762750 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d5aa02cd227e6ed8c626647ccdd2216d7675977d771c05587558c0576db0fda"} err="failed to get container status \"1d5aa02cd227e6ed8c626647ccdd2216d7675977d771c05587558c0576db0fda\": rpc error: code = NotFound desc = could not find container \"1d5aa02cd227e6ed8c626647ccdd2216d7675977d771c05587558c0576db0fda\": container with ID starting with 1d5aa02cd227e6ed8c626647ccdd2216d7675977d771c05587558c0576db0fda not found: ID does not exist" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.797290 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c52880d-4bb5-4665-964c-b378edcb5ac9" path="/var/lib/kubelet/pods/7c52880d-4bb5-4665-964c-b378edcb5ac9/volumes" Dec 10 19:11:19 crc kubenswrapper[4828]: I1210 19:11:19.798120 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3391181-ec59-42b3-b7a7-8f023bb54b9d" path="/var/lib/kubelet/pods/d3391181-ec59-42b3-b7a7-8f023bb54b9d/volumes" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.230343 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.230710 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.826362 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-crfqj"] Dec 10 19:11:21 crc kubenswrapper[4828]: E1210 19:11:21.826739 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3391181-ec59-42b3-b7a7-8f023bb54b9d" containerName="extract-utilities" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.826766 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3391181-ec59-42b3-b7a7-8f023bb54b9d" containerName="extract-utilities" Dec 10 19:11:21 crc kubenswrapper[4828]: E1210 19:11:21.826785 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3391181-ec59-42b3-b7a7-8f023bb54b9d" containerName="registry-server" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.826812 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3391181-ec59-42b3-b7a7-8f023bb54b9d" containerName="registry-server" Dec 10 19:11:21 crc kubenswrapper[4828]: E1210 19:11:21.826840 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c52880d-4bb5-4665-964c-b378edcb5ac9" containerName="extract-utilities" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.826850 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c52880d-4bb5-4665-964c-b378edcb5ac9" containerName="extract-utilities" Dec 10 19:11:21 crc kubenswrapper[4828]: E1210 19:11:21.826866 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c52880d-4bb5-4665-964c-b378edcb5ac9" containerName="extract-content" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.826875 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c52880d-4bb5-4665-964c-b378edcb5ac9" containerName="extract-content" Dec 10 19:11:21 crc kubenswrapper[4828]: E1210 19:11:21.826892 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c52880d-4bb5-4665-964c-b378edcb5ac9" containerName="registry-server" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.826900 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c52880d-4bb5-4665-964c-b378edcb5ac9" containerName="registry-server" Dec 10 19:11:21 crc kubenswrapper[4828]: E1210 19:11:21.826915 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3391181-ec59-42b3-b7a7-8f023bb54b9d" containerName="extract-content" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.826923 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3391181-ec59-42b3-b7a7-8f023bb54b9d" containerName="extract-content" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.827078 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c52880d-4bb5-4665-964c-b378edcb5ac9" containerName="registry-server" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.827101 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3391181-ec59-42b3-b7a7-8f023bb54b9d" containerName="registry-server" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.827772 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-crfqj" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.829596 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.830325 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-kxd9h" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.832040 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.832127 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.838248 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-crfqj"] Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.839203 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.843907 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.966013 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-crfqj"] Dec 10 19:11:21 crc kubenswrapper[4828]: E1210 19:11:21.966518 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver collector-token config config-openshift-service-cacrt datadir entrypoint kube-api-access-ff5p7 metrics sa-token tmp trusted-ca], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-logging/collector-crfqj" podUID="b67f9f6b-389c-4d57-927f-39b0828d5f61" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.967022 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-trusted-ca\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.967057 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff5p7\" (UniqueName: \"kubernetes.io/projected/b67f9f6b-389c-4d57-927f-39b0828d5f61-kube-api-access-ff5p7\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.967085 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-collector-token\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.967162 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/b67f9f6b-389c-4d57-927f-39b0828d5f61-tmp\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.967202 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-metrics\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.967306 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/b67f9f6b-389c-4d57-927f-39b0828d5f61-datadir\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.967327 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-entrypoint\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.967350 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-config-openshift-service-cacrt\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.967372 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/b67f9f6b-389c-4d57-927f-39b0828d5f61-sa-token\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.967402 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-collector-syslog-receiver\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:21 crc kubenswrapper[4828]: I1210 19:11:21.967440 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-config\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.069305 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-config\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.069363 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-trusted-ca\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.069387 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff5p7\" (UniqueName: \"kubernetes.io/projected/b67f9f6b-389c-4d57-927f-39b0828d5f61-kube-api-access-ff5p7\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.069405 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-collector-token\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.069434 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/b67f9f6b-389c-4d57-927f-39b0828d5f61-tmp\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.069466 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-metrics\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.069518 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/b67f9f6b-389c-4d57-927f-39b0828d5f61-datadir\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.069533 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-entrypoint\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.069548 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-config-openshift-service-cacrt\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.069567 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/b67f9f6b-389c-4d57-927f-39b0828d5f61-sa-token\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.069596 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-collector-syslog-receiver\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.070341 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/b67f9f6b-389c-4d57-927f-39b0828d5f61-datadir\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: E1210 19:11:22.070575 4828 secret.go:188] Couldn't get secret openshift-logging/collector-syslog-receiver: secret "collector-syslog-receiver" not found Dec 10 19:11:22 crc kubenswrapper[4828]: E1210 19:11:22.070631 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-collector-syslog-receiver podName:b67f9f6b-389c-4d57-927f-39b0828d5f61 nodeName:}" failed. No retries permitted until 2025-12-10 19:11:22.570616581 +0000 UTC m=+963.081227586 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "collector-syslog-receiver" (UniqueName: "kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-collector-syslog-receiver") pod "collector-crfqj" (UID: "b67f9f6b-389c-4d57-927f-39b0828d5f61") : secret "collector-syslog-receiver" not found Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.070673 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-config\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.070721 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-config-openshift-service-cacrt\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.070737 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-entrypoint\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.071099 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-trusted-ca\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.075191 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-metrics\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.079878 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/b67f9f6b-389c-4d57-927f-39b0828d5f61-tmp\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.080067 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-collector-token\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.089188 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/b67f9f6b-389c-4d57-927f-39b0828d5f61-sa-token\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.090983 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff5p7\" (UniqueName: \"kubernetes.io/projected/b67f9f6b-389c-4d57-927f-39b0828d5f61-kube-api-access-ff5p7\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.576735 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-collector-syslog-receiver\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.580044 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-collector-syslog-receiver\") pod \"collector-crfqj\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.678546 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.687822 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-crfqj" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.778618 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/b67f9f6b-389c-4d57-927f-39b0828d5f61-datadir\") pod \"b67f9f6b-389c-4d57-927f-39b0828d5f61\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.778848 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-metrics\") pod \"b67f9f6b-389c-4d57-927f-39b0828d5f61\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.778953 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-config-openshift-service-cacrt\") pod \"b67f9f6b-389c-4d57-927f-39b0828d5f61\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.778740 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b67f9f6b-389c-4d57-927f-39b0828d5f61-datadir" (OuterVolumeSpecName: "datadir") pod "b67f9f6b-389c-4d57-927f-39b0828d5f61" (UID: "b67f9f6b-389c-4d57-927f-39b0828d5f61"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.779115 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-collector-token\") pod \"b67f9f6b-389c-4d57-927f-39b0828d5f61\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.779211 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-config\") pod \"b67f9f6b-389c-4d57-927f-39b0828d5f61\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.779337 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-entrypoint\") pod \"b67f9f6b-389c-4d57-927f-39b0828d5f61\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.779461 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-trusted-ca\") pod \"b67f9f6b-389c-4d57-927f-39b0828d5f61\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.779556 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-collector-syslog-receiver\") pod \"b67f9f6b-389c-4d57-927f-39b0828d5f61\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.779639 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/b67f9f6b-389c-4d57-927f-39b0828d5f61-tmp\") pod \"b67f9f6b-389c-4d57-927f-39b0828d5f61\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.779690 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-config" (OuterVolumeSpecName: "config") pod "b67f9f6b-389c-4d57-927f-39b0828d5f61" (UID: "b67f9f6b-389c-4d57-927f-39b0828d5f61"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.779719 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "b67f9f6b-389c-4d57-927f-39b0828d5f61" (UID: "b67f9f6b-389c-4d57-927f-39b0828d5f61"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.779825 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/b67f9f6b-389c-4d57-927f-39b0828d5f61-sa-token\") pod \"b67f9f6b-389c-4d57-927f-39b0828d5f61\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.779910 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff5p7\" (UniqueName: \"kubernetes.io/projected/b67f9f6b-389c-4d57-927f-39b0828d5f61-kube-api-access-ff5p7\") pod \"b67f9f6b-389c-4d57-927f-39b0828d5f61\" (UID: \"b67f9f6b-389c-4d57-927f-39b0828d5f61\") " Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.779920 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "b67f9f6b-389c-4d57-927f-39b0828d5f61" (UID: "b67f9f6b-389c-4d57-927f-39b0828d5f61"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.780178 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "b67f9f6b-389c-4d57-927f-39b0828d5f61" (UID: "b67f9f6b-389c-4d57-927f-39b0828d5f61"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.780426 4828 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/b67f9f6b-389c-4d57-927f-39b0828d5f61-datadir\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.780523 4828 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.780595 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.780653 4828 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-entrypoint\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.780713 4828 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b67f9f6b-389c-4d57-927f-39b0828d5f61-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.781811 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-collector-token" (OuterVolumeSpecName: "collector-token") pod "b67f9f6b-389c-4d57-927f-39b0828d5f61" (UID: "b67f9f6b-389c-4d57-927f-39b0828d5f61"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.782325 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-metrics" (OuterVolumeSpecName: "metrics") pod "b67f9f6b-389c-4d57-927f-39b0828d5f61" (UID: "b67f9f6b-389c-4d57-927f-39b0828d5f61"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.782349 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "b67f9f6b-389c-4d57-927f-39b0828d5f61" (UID: "b67f9f6b-389c-4d57-927f-39b0828d5f61"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.782360 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b67f9f6b-389c-4d57-927f-39b0828d5f61-tmp" (OuterVolumeSpecName: "tmp") pod "b67f9f6b-389c-4d57-927f-39b0828d5f61" (UID: "b67f9f6b-389c-4d57-927f-39b0828d5f61"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.782745 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b67f9f6b-389c-4d57-927f-39b0828d5f61-kube-api-access-ff5p7" (OuterVolumeSpecName: "kube-api-access-ff5p7") pod "b67f9f6b-389c-4d57-927f-39b0828d5f61" (UID: "b67f9f6b-389c-4d57-927f-39b0828d5f61"). InnerVolumeSpecName "kube-api-access-ff5p7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.782879 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b67f9f6b-389c-4d57-927f-39b0828d5f61-sa-token" (OuterVolumeSpecName: "sa-token") pod "b67f9f6b-389c-4d57-927f-39b0828d5f61" (UID: "b67f9f6b-389c-4d57-927f-39b0828d5f61"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.883331 4828 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.883381 4828 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/b67f9f6b-389c-4d57-927f-39b0828d5f61-tmp\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.883400 4828 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/b67f9f6b-389c-4d57-927f-39b0828d5f61-sa-token\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.883418 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff5p7\" (UniqueName: \"kubernetes.io/projected/b67f9f6b-389c-4d57-927f-39b0828d5f61-kube-api-access-ff5p7\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.883435 4828 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-metrics\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:22 crc kubenswrapper[4828]: I1210 19:11:22.883452 4828 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/b67f9f6b-389c-4d57-927f-39b0828d5f61-collector-token\") on node \"crc\" DevicePath \"\"" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.693207 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-crfqj" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.766332 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-crfqj"] Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.785175 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-crfqj"] Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.802686 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b67f9f6b-389c-4d57-927f-39b0828d5f61" path="/var/lib/kubelet/pods/b67f9f6b-389c-4d57-927f-39b0828d5f61/volumes" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.802993 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-bwvf2"] Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.803823 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-bwvf2" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.809776 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.810654 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-kxd9h" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.810750 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.810683 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.810715 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.822374 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.834385 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-bwvf2"] Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.899672 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/416a301b-b484-4f47-9901-6c033705c501-collector-syslog-receiver\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.899969 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/416a301b-b484-4f47-9901-6c033705c501-tmp\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.900101 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/416a301b-b484-4f47-9901-6c033705c501-entrypoint\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.900192 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/416a301b-b484-4f47-9901-6c033705c501-config\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.900275 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/416a301b-b484-4f47-9901-6c033705c501-metrics\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.900370 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/416a301b-b484-4f47-9901-6c033705c501-collector-token\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.900460 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/416a301b-b484-4f47-9901-6c033705c501-config-openshift-service-cacrt\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.900561 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2m2d\" (UniqueName: \"kubernetes.io/projected/416a301b-b484-4f47-9901-6c033705c501-kube-api-access-m2m2d\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.900650 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/416a301b-b484-4f47-9901-6c033705c501-datadir\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.900771 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/416a301b-b484-4f47-9901-6c033705c501-trusted-ca\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:23 crc kubenswrapper[4828]: I1210 19:11:23.900938 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/416a301b-b484-4f47-9901-6c033705c501-sa-token\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.002484 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/416a301b-b484-4f47-9901-6c033705c501-tmp\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.002828 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/416a301b-b484-4f47-9901-6c033705c501-entrypoint\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.002958 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/416a301b-b484-4f47-9901-6c033705c501-config\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.003076 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/416a301b-b484-4f47-9901-6c033705c501-metrics\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.003193 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/416a301b-b484-4f47-9901-6c033705c501-collector-token\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.003303 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/416a301b-b484-4f47-9901-6c033705c501-config-openshift-service-cacrt\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.003395 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2m2d\" (UniqueName: \"kubernetes.io/projected/416a301b-b484-4f47-9901-6c033705c501-kube-api-access-m2m2d\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.003485 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/416a301b-b484-4f47-9901-6c033705c501-datadir\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.003611 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/416a301b-b484-4f47-9901-6c033705c501-trusted-ca\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.003717 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/416a301b-b484-4f47-9901-6c033705c501-sa-token\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.003863 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/416a301b-b484-4f47-9901-6c033705c501-collector-syslog-receiver\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.003783 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/416a301b-b484-4f47-9901-6c033705c501-config\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.004013 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/416a301b-b484-4f47-9901-6c033705c501-datadir\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.003777 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/416a301b-b484-4f47-9901-6c033705c501-entrypoint\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.005069 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/416a301b-b484-4f47-9901-6c033705c501-config-openshift-service-cacrt\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.006101 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/416a301b-b484-4f47-9901-6c033705c501-trusted-ca\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.008031 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/416a301b-b484-4f47-9901-6c033705c501-collector-syslog-receiver\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.009195 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/416a301b-b484-4f47-9901-6c033705c501-collector-token\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.010352 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/416a301b-b484-4f47-9901-6c033705c501-metrics\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.010546 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/416a301b-b484-4f47-9901-6c033705c501-tmp\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.024223 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/416a301b-b484-4f47-9901-6c033705c501-sa-token\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.024533 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2m2d\" (UniqueName: \"kubernetes.io/projected/416a301b-b484-4f47-9901-6c033705c501-kube-api-access-m2m2d\") pod \"collector-bwvf2\" (UID: \"416a301b-b484-4f47-9901-6c033705c501\") " pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.124670 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-bwvf2" Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.574710 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-bwvf2"] Dec 10 19:11:24 crc kubenswrapper[4828]: I1210 19:11:24.705566 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-bwvf2" event={"ID":"416a301b-b484-4f47-9901-6c033705c501","Type":"ContainerStarted","Data":"1b0d82886ac44760505e2e97b86177ff8742998368800ff608d24e516594eeda"} Dec 10 19:11:32 crc kubenswrapper[4828]: I1210 19:11:32.768883 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-bwvf2" event={"ID":"416a301b-b484-4f47-9901-6c033705c501","Type":"ContainerStarted","Data":"53ad408b4929d7a83217b7c1a3cd864cd17aa6208c9c148f3ad236b5808e4756"} Dec 10 19:11:32 crc kubenswrapper[4828]: I1210 19:11:32.797069 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-bwvf2" podStartSLOduration=2.613133116 podStartE2EDuration="9.797048849s" podCreationTimestamp="2025-12-10 19:11:23 +0000 UTC" firstStartedPulling="2025-12-10 19:11:24.585084205 +0000 UTC m=+965.095695220" lastFinishedPulling="2025-12-10 19:11:31.768999948 +0000 UTC m=+972.279610953" observedRunningTime="2025-12-10 19:11:32.791990907 +0000 UTC m=+973.302601942" watchObservedRunningTime="2025-12-10 19:11:32.797048849 +0000 UTC m=+973.307659864" Dec 10 19:11:51 crc kubenswrapper[4828]: I1210 19:11:51.230300 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:11:51 crc kubenswrapper[4828]: I1210 19:11:51.231008 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:11:51 crc kubenswrapper[4828]: I1210 19:11:51.231053 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 19:11:51 crc kubenswrapper[4828]: I1210 19:11:51.231756 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"97c29c3bfdeed7777dc023da2b04e9fc93587d77035f3f054b943162232fe57b"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:11:51 crc kubenswrapper[4828]: I1210 19:11:51.231822 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://97c29c3bfdeed7777dc023da2b04e9fc93587d77035f3f054b943162232fe57b" gracePeriod=600 Dec 10 19:11:51 crc kubenswrapper[4828]: I1210 19:11:51.912366 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="97c29c3bfdeed7777dc023da2b04e9fc93587d77035f3f054b943162232fe57b" exitCode=0 Dec 10 19:11:51 crc kubenswrapper[4828]: I1210 19:11:51.912568 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"97c29c3bfdeed7777dc023da2b04e9fc93587d77035f3f054b943162232fe57b"} Dec 10 19:11:51 crc kubenswrapper[4828]: I1210 19:11:51.912936 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"6e3402faca982d603d83c3dbbde8307326fa3a18260295ae4bfb613477d32c1e"} Dec 10 19:11:51 crc kubenswrapper[4828]: I1210 19:11:51.912959 4828 scope.go:117] "RemoveContainer" containerID="1d400e5fe2702c739b92ea52b2e9453f005b08c30f146bdcec9d4d95ba2cb93b" Dec 10 19:11:59 crc kubenswrapper[4828]: I1210 19:11:59.596655 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd"] Dec 10 19:11:59 crc kubenswrapper[4828]: I1210 19:11:59.598851 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" Dec 10 19:11:59 crc kubenswrapper[4828]: I1210 19:11:59.603863 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd"] Dec 10 19:11:59 crc kubenswrapper[4828]: I1210 19:11:59.606693 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 10 19:11:59 crc kubenswrapper[4828]: I1210 19:11:59.685108 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4knhf\" (UniqueName: \"kubernetes.io/projected/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-kube-api-access-4knhf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd\" (UID: \"7d03ca8f-a473-43a1-a97e-9b141ab54ee3\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" Dec 10 19:11:59 crc kubenswrapper[4828]: I1210 19:11:59.685186 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd\" (UID: \"7d03ca8f-a473-43a1-a97e-9b141ab54ee3\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" Dec 10 19:11:59 crc kubenswrapper[4828]: I1210 19:11:59.685284 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd\" (UID: \"7d03ca8f-a473-43a1-a97e-9b141ab54ee3\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" Dec 10 19:11:59 crc kubenswrapper[4828]: I1210 19:11:59.786415 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd\" (UID: \"7d03ca8f-a473-43a1-a97e-9b141ab54ee3\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" Dec 10 19:11:59 crc kubenswrapper[4828]: I1210 19:11:59.786692 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd\" (UID: \"7d03ca8f-a473-43a1-a97e-9b141ab54ee3\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" Dec 10 19:11:59 crc kubenswrapper[4828]: I1210 19:11:59.786773 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4knhf\" (UniqueName: \"kubernetes.io/projected/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-kube-api-access-4knhf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd\" (UID: \"7d03ca8f-a473-43a1-a97e-9b141ab54ee3\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" Dec 10 19:11:59 crc kubenswrapper[4828]: I1210 19:11:59.787530 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd\" (UID: \"7d03ca8f-a473-43a1-a97e-9b141ab54ee3\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" Dec 10 19:11:59 crc kubenswrapper[4828]: I1210 19:11:59.787709 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd\" (UID: \"7d03ca8f-a473-43a1-a97e-9b141ab54ee3\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" Dec 10 19:11:59 crc kubenswrapper[4828]: I1210 19:11:59.804459 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4knhf\" (UniqueName: \"kubernetes.io/projected/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-kube-api-access-4knhf\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd\" (UID: \"7d03ca8f-a473-43a1-a97e-9b141ab54ee3\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" Dec 10 19:11:59 crc kubenswrapper[4828]: I1210 19:11:59.919822 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" Dec 10 19:12:00 crc kubenswrapper[4828]: I1210 19:12:00.371126 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd"] Dec 10 19:12:00 crc kubenswrapper[4828]: I1210 19:12:00.994197 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" event={"ID":"7d03ca8f-a473-43a1-a97e-9b141ab54ee3","Type":"ContainerStarted","Data":"4bbe2c013246488187ab29a2760465b749346b8276c1346adb71375876456d7d"} Dec 10 19:12:02 crc kubenswrapper[4828]: I1210 19:12:02.005201 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" event={"ID":"7d03ca8f-a473-43a1-a97e-9b141ab54ee3","Type":"ContainerStarted","Data":"cb4c352410d21a888508b3c76bfd7ceb4d9813bee91d7a19f10b0dc348691e9f"} Dec 10 19:12:03 crc kubenswrapper[4828]: I1210 19:12:03.013645 4828 generic.go:334] "Generic (PLEG): container finished" podID="7d03ca8f-a473-43a1-a97e-9b141ab54ee3" containerID="cb4c352410d21a888508b3c76bfd7ceb4d9813bee91d7a19f10b0dc348691e9f" exitCode=0 Dec 10 19:12:03 crc kubenswrapper[4828]: I1210 19:12:03.013686 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" event={"ID":"7d03ca8f-a473-43a1-a97e-9b141ab54ee3","Type":"ContainerDied","Data":"cb4c352410d21a888508b3c76bfd7ceb4d9813bee91d7a19f10b0dc348691e9f"} Dec 10 19:12:05 crc kubenswrapper[4828]: I1210 19:12:05.029419 4828 generic.go:334] "Generic (PLEG): container finished" podID="7d03ca8f-a473-43a1-a97e-9b141ab54ee3" containerID="b74c26a4a3aafd3bc8c0f88647f395c7f2b3e8ca52b4b0d31e2a3614842d5a16" exitCode=0 Dec 10 19:12:05 crc kubenswrapper[4828]: I1210 19:12:05.029505 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" event={"ID":"7d03ca8f-a473-43a1-a97e-9b141ab54ee3","Type":"ContainerDied","Data":"b74c26a4a3aafd3bc8c0f88647f395c7f2b3e8ca52b4b0d31e2a3614842d5a16"} Dec 10 19:12:06 crc kubenswrapper[4828]: I1210 19:12:06.038046 4828 generic.go:334] "Generic (PLEG): container finished" podID="7d03ca8f-a473-43a1-a97e-9b141ab54ee3" containerID="bc5fc48fa6010bcb3f5e39b2e4589c7374cb9de27edd73f4a5f838be1ee1d0d7" exitCode=0 Dec 10 19:12:06 crc kubenswrapper[4828]: I1210 19:12:06.038094 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" event={"ID":"7d03ca8f-a473-43a1-a97e-9b141ab54ee3","Type":"ContainerDied","Data":"bc5fc48fa6010bcb3f5e39b2e4589c7374cb9de27edd73f4a5f838be1ee1d0d7"} Dec 10 19:12:07 crc kubenswrapper[4828]: I1210 19:12:07.424664 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" Dec 10 19:12:07 crc kubenswrapper[4828]: I1210 19:12:07.602283 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-bundle\") pod \"7d03ca8f-a473-43a1-a97e-9b141ab54ee3\" (UID: \"7d03ca8f-a473-43a1-a97e-9b141ab54ee3\") " Dec 10 19:12:07 crc kubenswrapper[4828]: I1210 19:12:07.602354 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4knhf\" (UniqueName: \"kubernetes.io/projected/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-kube-api-access-4knhf\") pod \"7d03ca8f-a473-43a1-a97e-9b141ab54ee3\" (UID: \"7d03ca8f-a473-43a1-a97e-9b141ab54ee3\") " Dec 10 19:12:07 crc kubenswrapper[4828]: I1210 19:12:07.602463 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-util\") pod \"7d03ca8f-a473-43a1-a97e-9b141ab54ee3\" (UID: \"7d03ca8f-a473-43a1-a97e-9b141ab54ee3\") " Dec 10 19:12:07 crc kubenswrapper[4828]: I1210 19:12:07.603199 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-bundle" (OuterVolumeSpecName: "bundle") pod "7d03ca8f-a473-43a1-a97e-9b141ab54ee3" (UID: "7d03ca8f-a473-43a1-a97e-9b141ab54ee3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:12:07 crc kubenswrapper[4828]: I1210 19:12:07.603826 4828 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:12:07 crc kubenswrapper[4828]: I1210 19:12:07.608694 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-kube-api-access-4knhf" (OuterVolumeSpecName: "kube-api-access-4knhf") pod "7d03ca8f-a473-43a1-a97e-9b141ab54ee3" (UID: "7d03ca8f-a473-43a1-a97e-9b141ab54ee3"). InnerVolumeSpecName "kube-api-access-4knhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:12:07 crc kubenswrapper[4828]: I1210 19:12:07.705843 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4knhf\" (UniqueName: \"kubernetes.io/projected/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-kube-api-access-4knhf\") on node \"crc\" DevicePath \"\"" Dec 10 19:12:07 crc kubenswrapper[4828]: I1210 19:12:07.719318 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-util" (OuterVolumeSpecName: "util") pod "7d03ca8f-a473-43a1-a97e-9b141ab54ee3" (UID: "7d03ca8f-a473-43a1-a97e-9b141ab54ee3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:12:07 crc kubenswrapper[4828]: I1210 19:12:07.807825 4828 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7d03ca8f-a473-43a1-a97e-9b141ab54ee3-util\") on node \"crc\" DevicePath \"\"" Dec 10 19:12:08 crc kubenswrapper[4828]: I1210 19:12:08.053991 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" event={"ID":"7d03ca8f-a473-43a1-a97e-9b141ab54ee3","Type":"ContainerDied","Data":"4bbe2c013246488187ab29a2760465b749346b8276c1346adb71375876456d7d"} Dec 10 19:12:08 crc kubenswrapper[4828]: I1210 19:12:08.054032 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bbe2c013246488187ab29a2760465b749346b8276c1346adb71375876456d7d" Dec 10 19:12:08 crc kubenswrapper[4828]: I1210 19:12:08.054127 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd" Dec 10 19:12:11 crc kubenswrapper[4828]: I1210 19:12:11.357223 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-d4x88"] Dec 10 19:12:11 crc kubenswrapper[4828]: E1210 19:12:11.359254 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d03ca8f-a473-43a1-a97e-9b141ab54ee3" containerName="util" Dec 10 19:12:11 crc kubenswrapper[4828]: I1210 19:12:11.359356 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d03ca8f-a473-43a1-a97e-9b141ab54ee3" containerName="util" Dec 10 19:12:11 crc kubenswrapper[4828]: E1210 19:12:11.359476 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d03ca8f-a473-43a1-a97e-9b141ab54ee3" containerName="extract" Dec 10 19:12:11 crc kubenswrapper[4828]: I1210 19:12:11.359571 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d03ca8f-a473-43a1-a97e-9b141ab54ee3" containerName="extract" Dec 10 19:12:11 crc kubenswrapper[4828]: E1210 19:12:11.359673 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d03ca8f-a473-43a1-a97e-9b141ab54ee3" containerName="pull" Dec 10 19:12:11 crc kubenswrapper[4828]: I1210 19:12:11.359751 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d03ca8f-a473-43a1-a97e-9b141ab54ee3" containerName="pull" Dec 10 19:12:11 crc kubenswrapper[4828]: I1210 19:12:11.360052 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d03ca8f-a473-43a1-a97e-9b141ab54ee3" containerName="extract" Dec 10 19:12:11 crc kubenswrapper[4828]: I1210 19:12:11.360901 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-d4x88" Dec 10 19:12:11 crc kubenswrapper[4828]: I1210 19:12:11.365470 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 10 19:12:11 crc kubenswrapper[4828]: I1210 19:12:11.365569 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-c67qh" Dec 10 19:12:11 crc kubenswrapper[4828]: I1210 19:12:11.365628 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 10 19:12:11 crc kubenswrapper[4828]: I1210 19:12:11.366422 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-d4x88"] Dec 10 19:12:11 crc kubenswrapper[4828]: I1210 19:12:11.562503 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t8n5\" (UniqueName: \"kubernetes.io/projected/9c4e3c85-4f86-434a-8adb-d22d17b3b44c-kube-api-access-8t8n5\") pod \"nmstate-operator-5b5b58f5c8-d4x88\" (UID: \"9c4e3c85-4f86-434a-8adb-d22d17b3b44c\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-d4x88" Dec 10 19:12:11 crc kubenswrapper[4828]: I1210 19:12:11.664069 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t8n5\" (UniqueName: \"kubernetes.io/projected/9c4e3c85-4f86-434a-8adb-d22d17b3b44c-kube-api-access-8t8n5\") pod \"nmstate-operator-5b5b58f5c8-d4x88\" (UID: \"9c4e3c85-4f86-434a-8adb-d22d17b3b44c\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-d4x88" Dec 10 19:12:11 crc kubenswrapper[4828]: I1210 19:12:11.689901 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t8n5\" (UniqueName: \"kubernetes.io/projected/9c4e3c85-4f86-434a-8adb-d22d17b3b44c-kube-api-access-8t8n5\") pod \"nmstate-operator-5b5b58f5c8-d4x88\" (UID: \"9c4e3c85-4f86-434a-8adb-d22d17b3b44c\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-d4x88" Dec 10 19:12:11 crc kubenswrapper[4828]: I1210 19:12:11.981549 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-d4x88" Dec 10 19:12:12 crc kubenswrapper[4828]: I1210 19:12:12.432602 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-d4x88"] Dec 10 19:12:12 crc kubenswrapper[4828]: W1210 19:12:12.436530 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c4e3c85_4f86_434a_8adb_d22d17b3b44c.slice/crio-289fe9b5d1ca10b5314db2f3b075272400d1a055c25e9d29c09ca1783bdeda45 WatchSource:0}: Error finding container 289fe9b5d1ca10b5314db2f3b075272400d1a055c25e9d29c09ca1783bdeda45: Status 404 returned error can't find the container with id 289fe9b5d1ca10b5314db2f3b075272400d1a055c25e9d29c09ca1783bdeda45 Dec 10 19:12:12 crc kubenswrapper[4828]: I1210 19:12:12.438786 4828 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:12:13 crc kubenswrapper[4828]: I1210 19:12:13.087858 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-d4x88" event={"ID":"9c4e3c85-4f86-434a-8adb-d22d17b3b44c","Type":"ContainerStarted","Data":"289fe9b5d1ca10b5314db2f3b075272400d1a055c25e9d29c09ca1783bdeda45"} Dec 10 19:12:15 crc kubenswrapper[4828]: I1210 19:12:15.103786 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-d4x88" event={"ID":"9c4e3c85-4f86-434a-8adb-d22d17b3b44c","Type":"ContainerStarted","Data":"0f2efd140dd99c731d7157b01a0f269569536b98c62a36c93f25ce84630b27f7"} Dec 10 19:12:15 crc kubenswrapper[4828]: I1210 19:12:15.121323 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-d4x88" podStartSLOduration=1.8147858860000001 podStartE2EDuration="4.121299371s" podCreationTimestamp="2025-12-10 19:12:11 +0000 UTC" firstStartedPulling="2025-12-10 19:12:12.438513868 +0000 UTC m=+1012.949124873" lastFinishedPulling="2025-12-10 19:12:14.745027353 +0000 UTC m=+1015.255638358" observedRunningTime="2025-12-10 19:12:15.118639741 +0000 UTC m=+1015.629250766" watchObservedRunningTime="2025-12-10 19:12:15.121299371 +0000 UTC m=+1015.631910376" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.663333 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-rz5wn"] Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.665287 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-rz5wn" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.667199 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-vm9sc" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.671457 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j9tv4"] Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.672628 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j9tv4" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.674481 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.680430 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j9tv4"] Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.691175 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-rz5wn"] Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.698692 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnwvx\" (UniqueName: \"kubernetes.io/projected/afc9bf44-dee0-4c71-8e9a-519f382d1857-kube-api-access-nnwvx\") pod \"nmstate-webhook-5f6d4c5ccb-j9tv4\" (UID: \"afc9bf44-dee0-4c71-8e9a-519f382d1857\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j9tv4" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.698985 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/afc9bf44-dee0-4c71-8e9a-519f382d1857-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-j9tv4\" (UID: \"afc9bf44-dee0-4c71-8e9a-519f382d1857\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j9tv4" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.699146 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdjwk\" (UniqueName: \"kubernetes.io/projected/9f6baf7f-f16b-4acf-a2c5-18d72615ff42-kube-api-access-vdjwk\") pod \"nmstate-metrics-7f946cbc9-rz5wn\" (UID: \"9f6baf7f-f16b-4acf-a2c5-18d72615ff42\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-rz5wn" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.700003 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-5tn4b"] Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.701501 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-5tn4b" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.811523 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnwvx\" (UniqueName: \"kubernetes.io/projected/afc9bf44-dee0-4c71-8e9a-519f382d1857-kube-api-access-nnwvx\") pod \"nmstate-webhook-5f6d4c5ccb-j9tv4\" (UID: \"afc9bf44-dee0-4c71-8e9a-519f382d1857\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j9tv4" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.817078 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/afc9bf44-dee0-4c71-8e9a-519f382d1857-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-j9tv4\" (UID: \"afc9bf44-dee0-4c71-8e9a-519f382d1857\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j9tv4" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.818569 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdjwk\" (UniqueName: \"kubernetes.io/projected/9f6baf7f-f16b-4acf-a2c5-18d72615ff42-kube-api-access-vdjwk\") pod \"nmstate-metrics-7f946cbc9-rz5wn\" (UID: \"9f6baf7f-f16b-4acf-a2c5-18d72615ff42\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-rz5wn" Dec 10 19:12:20 crc kubenswrapper[4828]: E1210 19:12:20.818946 4828 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 10 19:12:20 crc kubenswrapper[4828]: E1210 19:12:20.819030 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/afc9bf44-dee0-4c71-8e9a-519f382d1857-tls-key-pair podName:afc9bf44-dee0-4c71-8e9a-519f382d1857 nodeName:}" failed. No retries permitted until 2025-12-10 19:12:21.319011096 +0000 UTC m=+1021.829622101 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/afc9bf44-dee0-4c71-8e9a-519f382d1857-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-j9tv4" (UID: "afc9bf44-dee0-4c71-8e9a-519f382d1857") : secret "openshift-nmstate-webhook" not found Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.842498 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnwvx\" (UniqueName: \"kubernetes.io/projected/afc9bf44-dee0-4c71-8e9a-519f382d1857-kube-api-access-nnwvx\") pod \"nmstate-webhook-5f6d4c5ccb-j9tv4\" (UID: \"afc9bf44-dee0-4c71-8e9a-519f382d1857\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j9tv4" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.842758 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r"] Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.843876 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdjwk\" (UniqueName: \"kubernetes.io/projected/9f6baf7f-f16b-4acf-a2c5-18d72615ff42-kube-api-access-vdjwk\") pod \"nmstate-metrics-7f946cbc9-rz5wn\" (UID: \"9f6baf7f-f16b-4acf-a2c5-18d72615ff42\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-rz5wn" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.844445 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.849079 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-sdld5" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.849690 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.849891 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.863625 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r"] Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.921410 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpmth\" (UniqueName: \"kubernetes.io/projected/ce59be34-eb49-4a48-b2cb-e627bf2ed853-kube-api-access-tpmth\") pod \"nmstate-handler-5tn4b\" (UID: \"ce59be34-eb49-4a48-b2cb-e627bf2ed853\") " pod="openshift-nmstate/nmstate-handler-5tn4b" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.921481 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ce59be34-eb49-4a48-b2cb-e627bf2ed853-nmstate-lock\") pod \"nmstate-handler-5tn4b\" (UID: \"ce59be34-eb49-4a48-b2cb-e627bf2ed853\") " pod="openshift-nmstate/nmstate-handler-5tn4b" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.921556 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ce59be34-eb49-4a48-b2cb-e627bf2ed853-dbus-socket\") pod \"nmstate-handler-5tn4b\" (UID: \"ce59be34-eb49-4a48-b2cb-e627bf2ed853\") " pod="openshift-nmstate/nmstate-handler-5tn4b" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.921593 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6c267cf9-bc24-4e82-8293-464d90e3dbc2-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-bdm9r\" (UID: \"6c267cf9-bc24-4e82-8293-464d90e3dbc2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.921633 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ce59be34-eb49-4a48-b2cb-e627bf2ed853-ovs-socket\") pod \"nmstate-handler-5tn4b\" (UID: \"ce59be34-eb49-4a48-b2cb-e627bf2ed853\") " pod="openshift-nmstate/nmstate-handler-5tn4b" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.921654 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgqsb\" (UniqueName: \"kubernetes.io/projected/6c267cf9-bc24-4e82-8293-464d90e3dbc2-kube-api-access-kgqsb\") pod \"nmstate-console-plugin-7fbb5f6569-bdm9r\" (UID: \"6c267cf9-bc24-4e82-8293-464d90e3dbc2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.921696 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/6c267cf9-bc24-4e82-8293-464d90e3dbc2-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-bdm9r\" (UID: \"6c267cf9-bc24-4e82-8293-464d90e3dbc2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.991111 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-rz5wn" Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.991958 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7994d84d8-5xz2j"] Dec 10 19:12:20 crc kubenswrapper[4828]: I1210 19:12:20.993059 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.016194 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7994d84d8-5xz2j"] Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.023705 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ce59be34-eb49-4a48-b2cb-e627bf2ed853-dbus-socket\") pod \"nmstate-handler-5tn4b\" (UID: \"ce59be34-eb49-4a48-b2cb-e627bf2ed853\") " pod="openshift-nmstate/nmstate-handler-5tn4b" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.023755 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6c267cf9-bc24-4e82-8293-464d90e3dbc2-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-bdm9r\" (UID: \"6c267cf9-bc24-4e82-8293-464d90e3dbc2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.023781 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-serving-cert\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.023820 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-oauth-serving-cert\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.023840 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-trusted-ca-bundle\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.023862 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ce59be34-eb49-4a48-b2cb-e627bf2ed853-ovs-socket\") pod \"nmstate-handler-5tn4b\" (UID: \"ce59be34-eb49-4a48-b2cb-e627bf2ed853\") " pod="openshift-nmstate/nmstate-handler-5tn4b" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.023881 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgqsb\" (UniqueName: \"kubernetes.io/projected/6c267cf9-bc24-4e82-8293-464d90e3dbc2-kube-api-access-kgqsb\") pod \"nmstate-console-plugin-7fbb5f6569-bdm9r\" (UID: \"6c267cf9-bc24-4e82-8293-464d90e3dbc2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.023900 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72rch\" (UniqueName: \"kubernetes.io/projected/5bd7de50-96a2-4ead-9ea2-f891902f61f1-kube-api-access-72rch\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.023935 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/6c267cf9-bc24-4e82-8293-464d90e3dbc2-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-bdm9r\" (UID: \"6c267cf9-bc24-4e82-8293-464d90e3dbc2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.023955 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-config\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.023928 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ce59be34-eb49-4a48-b2cb-e627bf2ed853-dbus-socket\") pod \"nmstate-handler-5tn4b\" (UID: \"ce59be34-eb49-4a48-b2cb-e627bf2ed853\") " pod="openshift-nmstate/nmstate-handler-5tn4b" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.024052 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-service-ca\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: E1210 19:12:21.024954 4828 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 10 19:12:21 crc kubenswrapper[4828]: E1210 19:12:21.025023 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6c267cf9-bc24-4e82-8293-464d90e3dbc2-plugin-serving-cert podName:6c267cf9-bc24-4e82-8293-464d90e3dbc2 nodeName:}" failed. No retries permitted until 2025-12-10 19:12:21.525001754 +0000 UTC m=+1022.035612819 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/6c267cf9-bc24-4e82-8293-464d90e3dbc2-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-bdm9r" (UID: "6c267cf9-bc24-4e82-8293-464d90e3dbc2") : secret "plugin-serving-cert" not found Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.025050 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6c267cf9-bc24-4e82-8293-464d90e3dbc2-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-bdm9r\" (UID: \"6c267cf9-bc24-4e82-8293-464d90e3dbc2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.025140 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpmth\" (UniqueName: \"kubernetes.io/projected/ce59be34-eb49-4a48-b2cb-e627bf2ed853-kube-api-access-tpmth\") pod \"nmstate-handler-5tn4b\" (UID: \"ce59be34-eb49-4a48-b2cb-e627bf2ed853\") " pod="openshift-nmstate/nmstate-handler-5tn4b" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.025162 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ce59be34-eb49-4a48-b2cb-e627bf2ed853-nmstate-lock\") pod \"nmstate-handler-5tn4b\" (UID: \"ce59be34-eb49-4a48-b2cb-e627bf2ed853\") " pod="openshift-nmstate/nmstate-handler-5tn4b" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.025167 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ce59be34-eb49-4a48-b2cb-e627bf2ed853-ovs-socket\") pod \"nmstate-handler-5tn4b\" (UID: \"ce59be34-eb49-4a48-b2cb-e627bf2ed853\") " pod="openshift-nmstate/nmstate-handler-5tn4b" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.025182 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-oauth-config\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.025252 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ce59be34-eb49-4a48-b2cb-e627bf2ed853-nmstate-lock\") pod \"nmstate-handler-5tn4b\" (UID: \"ce59be34-eb49-4a48-b2cb-e627bf2ed853\") " pod="openshift-nmstate/nmstate-handler-5tn4b" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.046317 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgqsb\" (UniqueName: \"kubernetes.io/projected/6c267cf9-bc24-4e82-8293-464d90e3dbc2-kube-api-access-kgqsb\") pod \"nmstate-console-plugin-7fbb5f6569-bdm9r\" (UID: \"6c267cf9-bc24-4e82-8293-464d90e3dbc2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.048945 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpmth\" (UniqueName: \"kubernetes.io/projected/ce59be34-eb49-4a48-b2cb-e627bf2ed853-kube-api-access-tpmth\") pod \"nmstate-handler-5tn4b\" (UID: \"ce59be34-eb49-4a48-b2cb-e627bf2ed853\") " pod="openshift-nmstate/nmstate-handler-5tn4b" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.128325 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-serving-cert\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.128643 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-oauth-serving-cert\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.128664 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-trusted-ca-bundle\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.128708 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72rch\" (UniqueName: \"kubernetes.io/projected/5bd7de50-96a2-4ead-9ea2-f891902f61f1-kube-api-access-72rch\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.128757 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-config\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.128775 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-service-ca\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.128873 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-oauth-config\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.130241 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-oauth-serving-cert\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.130438 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-trusted-ca-bundle\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.131048 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-config\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.135088 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-service-ca\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.135449 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-oauth-config\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.135688 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-serving-cert\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.154454 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72rch\" (UniqueName: \"kubernetes.io/projected/5bd7de50-96a2-4ead-9ea2-f891902f61f1-kube-api-access-72rch\") pod \"console-7994d84d8-5xz2j\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.257301 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-rz5wn"] Dec 10 19:12:21 crc kubenswrapper[4828]: W1210 19:12:21.258524 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f6baf7f_f16b_4acf_a2c5_18d72615ff42.slice/crio-dd4c495d331fee2df9929ba00aee701ed9887a35d782794186505f50ce6e6859 WatchSource:0}: Error finding container dd4c495d331fee2df9929ba00aee701ed9887a35d782794186505f50ce6e6859: Status 404 returned error can't find the container with id dd4c495d331fee2df9929ba00aee701ed9887a35d782794186505f50ce6e6859 Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.323615 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-5tn4b" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.332408 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/afc9bf44-dee0-4c71-8e9a-519f382d1857-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-j9tv4\" (UID: \"afc9bf44-dee0-4c71-8e9a-519f382d1857\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j9tv4" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.337225 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/afc9bf44-dee0-4c71-8e9a-519f382d1857-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-j9tv4\" (UID: \"afc9bf44-dee0-4c71-8e9a-519f382d1857\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j9tv4" Dec 10 19:12:21 crc kubenswrapper[4828]: W1210 19:12:21.346009 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce59be34_eb49_4a48_b2cb_e627bf2ed853.slice/crio-0b48174901250794b79f558e576a49d606853dd6c8d14edb659b98b15e2ce4d4 WatchSource:0}: Error finding container 0b48174901250794b79f558e576a49d606853dd6c8d14edb659b98b15e2ce4d4: Status 404 returned error can't find the container with id 0b48174901250794b79f558e576a49d606853dd6c8d14edb659b98b15e2ce4d4 Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.395793 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.535667 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/6c267cf9-bc24-4e82-8293-464d90e3dbc2-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-bdm9r\" (UID: \"6c267cf9-bc24-4e82-8293-464d90e3dbc2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.541429 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/6c267cf9-bc24-4e82-8293-464d90e3dbc2-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-bdm9r\" (UID: \"6c267cf9-bc24-4e82-8293-464d90e3dbc2\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.599352 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j9tv4" Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.656853 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7994d84d8-5xz2j"] Dec 10 19:12:21 crc kubenswrapper[4828]: W1210 19:12:21.664235 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5bd7de50_96a2_4ead_9ea2_f891902f61f1.slice/crio-a8925697c5ac2db98824d698d517c7a4ac5e9d33034b915c34fbfe7120225e35 WatchSource:0}: Error finding container a8925697c5ac2db98824d698d517c7a4ac5e9d33034b915c34fbfe7120225e35: Status 404 returned error can't find the container with id a8925697c5ac2db98824d698d517c7a4ac5e9d33034b915c34fbfe7120225e35 Dec 10 19:12:21 crc kubenswrapper[4828]: I1210 19:12:21.818139 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r" Dec 10 19:12:22 crc kubenswrapper[4828]: I1210 19:12:22.021252 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j9tv4"] Dec 10 19:12:22 crc kubenswrapper[4828]: I1210 19:12:22.156485 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-5tn4b" event={"ID":"ce59be34-eb49-4a48-b2cb-e627bf2ed853","Type":"ContainerStarted","Data":"0b48174901250794b79f558e576a49d606853dd6c8d14edb659b98b15e2ce4d4"} Dec 10 19:12:22 crc kubenswrapper[4828]: I1210 19:12:22.158084 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7994d84d8-5xz2j" event={"ID":"5bd7de50-96a2-4ead-9ea2-f891902f61f1","Type":"ContainerStarted","Data":"885ed7db92fb53103d12f67bbe8e1c5fa4606fb23015c598be39762944f115fd"} Dec 10 19:12:22 crc kubenswrapper[4828]: I1210 19:12:22.158133 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7994d84d8-5xz2j" event={"ID":"5bd7de50-96a2-4ead-9ea2-f891902f61f1","Type":"ContainerStarted","Data":"a8925697c5ac2db98824d698d517c7a4ac5e9d33034b915c34fbfe7120225e35"} Dec 10 19:12:22 crc kubenswrapper[4828]: I1210 19:12:22.159896 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j9tv4" event={"ID":"afc9bf44-dee0-4c71-8e9a-519f382d1857","Type":"ContainerStarted","Data":"07fb267a85ff7b809d4e461c5cf6ec6acbf5b35f3a76825c809e3d5a4e0e1181"} Dec 10 19:12:22 crc kubenswrapper[4828]: I1210 19:12:22.160758 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-rz5wn" event={"ID":"9f6baf7f-f16b-4acf-a2c5-18d72615ff42","Type":"ContainerStarted","Data":"dd4c495d331fee2df9929ba00aee701ed9887a35d782794186505f50ce6e6859"} Dec 10 19:12:22 crc kubenswrapper[4828]: I1210 19:12:22.204787 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7994d84d8-5xz2j" podStartSLOduration=2.204765302 podStartE2EDuration="2.204765302s" podCreationTimestamp="2025-12-10 19:12:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:12:22.177653507 +0000 UTC m=+1022.688264512" watchObservedRunningTime="2025-12-10 19:12:22.204765302 +0000 UTC m=+1022.715376327" Dec 10 19:12:22 crc kubenswrapper[4828]: I1210 19:12:22.206072 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r"] Dec 10 19:12:23 crc kubenswrapper[4828]: I1210 19:12:23.171967 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r" event={"ID":"6c267cf9-bc24-4e82-8293-464d90e3dbc2","Type":"ContainerStarted","Data":"cd1fa9a9f76ae2e21a8c363ff2e147b7e7b080b25d7e43af9e310fb544435f55"} Dec 10 19:12:25 crc kubenswrapper[4828]: I1210 19:12:25.199305 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j9tv4" event={"ID":"afc9bf44-dee0-4c71-8e9a-519f382d1857","Type":"ContainerStarted","Data":"a1981bdeb2ff3a4a51b9b3598034acffcdbbe72b840ec7009512ad3ae75881b4"} Dec 10 19:12:25 crc kubenswrapper[4828]: I1210 19:12:25.201036 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j9tv4" Dec 10 19:12:25 crc kubenswrapper[4828]: I1210 19:12:25.202126 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-rz5wn" event={"ID":"9f6baf7f-f16b-4acf-a2c5-18d72615ff42","Type":"ContainerStarted","Data":"6554e522f8f0417348111ed8646c44068511aad5e6c62b1ed603bbdd880a77e8"} Dec 10 19:12:25 crc kubenswrapper[4828]: I1210 19:12:25.203688 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-5tn4b" event={"ID":"ce59be34-eb49-4a48-b2cb-e627bf2ed853","Type":"ContainerStarted","Data":"6b13edd2faf9a1e7074ee6cdc2c33c0f917262c9b37f6fc2992a47cda4866c08"} Dec 10 19:12:25 crc kubenswrapper[4828]: I1210 19:12:25.204122 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-5tn4b" Dec 10 19:12:25 crc kubenswrapper[4828]: I1210 19:12:25.205511 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r" event={"ID":"6c267cf9-bc24-4e82-8293-464d90e3dbc2","Type":"ContainerStarted","Data":"d636f6e17df4cdfe46089f54793ff0c45c45c1add050ce9744f1d93018e599b1"} Dec 10 19:12:25 crc kubenswrapper[4828]: I1210 19:12:25.220101 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j9tv4" podStartSLOduration=2.418758668 podStartE2EDuration="5.220082674s" podCreationTimestamp="2025-12-10 19:12:20 +0000 UTC" firstStartedPulling="2025-12-10 19:12:22.025238462 +0000 UTC m=+1022.535849467" lastFinishedPulling="2025-12-10 19:12:24.826562468 +0000 UTC m=+1025.337173473" observedRunningTime="2025-12-10 19:12:25.21722113 +0000 UTC m=+1025.727832145" watchObservedRunningTime="2025-12-10 19:12:25.220082674 +0000 UTC m=+1025.730693679" Dec 10 19:12:25 crc kubenswrapper[4828]: I1210 19:12:25.245822 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-5tn4b" podStartSLOduration=1.7586744479999998 podStartE2EDuration="5.245785483s" podCreationTimestamp="2025-12-10 19:12:20 +0000 UTC" firstStartedPulling="2025-12-10 19:12:21.347738929 +0000 UTC m=+1021.858349944" lastFinishedPulling="2025-12-10 19:12:24.834849974 +0000 UTC m=+1025.345460979" observedRunningTime="2025-12-10 19:12:25.243962235 +0000 UTC m=+1025.754573250" watchObservedRunningTime="2025-12-10 19:12:25.245785483 +0000 UTC m=+1025.756396488" Dec 10 19:12:25 crc kubenswrapper[4828]: I1210 19:12:25.261737 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-bdm9r" podStartSLOduration=2.658977606 podStartE2EDuration="5.261719317s" podCreationTimestamp="2025-12-10 19:12:20 +0000 UTC" firstStartedPulling="2025-12-10 19:12:22.222316978 +0000 UTC m=+1022.732927983" lastFinishedPulling="2025-12-10 19:12:24.825058689 +0000 UTC m=+1025.335669694" observedRunningTime="2025-12-10 19:12:25.25683518 +0000 UTC m=+1025.767446185" watchObservedRunningTime="2025-12-10 19:12:25.261719317 +0000 UTC m=+1025.772330322" Dec 10 19:12:28 crc kubenswrapper[4828]: I1210 19:12:28.227773 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-rz5wn" event={"ID":"9f6baf7f-f16b-4acf-a2c5-18d72615ff42","Type":"ContainerStarted","Data":"2991609734edc8ef7cbf857a0338c9a622666bcf81309a3327b081d9ac434e49"} Dec 10 19:12:28 crc kubenswrapper[4828]: I1210 19:12:28.248461 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-rz5wn" podStartSLOduration=2.354121516 podStartE2EDuration="8.248437715s" podCreationTimestamp="2025-12-10 19:12:20 +0000 UTC" firstStartedPulling="2025-12-10 19:12:21.26010071 +0000 UTC m=+1021.770711715" lastFinishedPulling="2025-12-10 19:12:27.154416909 +0000 UTC m=+1027.665027914" observedRunningTime="2025-12-10 19:12:28.242562883 +0000 UTC m=+1028.753173898" watchObservedRunningTime="2025-12-10 19:12:28.248437715 +0000 UTC m=+1028.759048720" Dec 10 19:12:31 crc kubenswrapper[4828]: I1210 19:12:31.349867 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-5tn4b" Dec 10 19:12:31 crc kubenswrapper[4828]: I1210 19:12:31.396069 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:31 crc kubenswrapper[4828]: I1210 19:12:31.396135 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:31 crc kubenswrapper[4828]: I1210 19:12:31.403374 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:32 crc kubenswrapper[4828]: I1210 19:12:32.258837 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:12:32 crc kubenswrapper[4828]: I1210 19:12:32.324538 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-8486f475fd-6jrxr"] Dec 10 19:12:41 crc kubenswrapper[4828]: I1210 19:12:41.605565 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-j9tv4" Dec 10 19:12:57 crc kubenswrapper[4828]: I1210 19:12:57.394680 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-8486f475fd-6jrxr" podUID="434cf9e4-da1b-49cb-a641-9a84ebe2e22c" containerName="console" containerID="cri-o://2e35b4ad0b8cb77ea24d1d1476570efc4ac30280c569dc13dfaf10b670724d9b" gracePeriod=15 Dec 10 19:12:57 crc kubenswrapper[4828]: I1210 19:12:57.895775 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-8486f475fd-6jrxr_434cf9e4-da1b-49cb-a641-9a84ebe2e22c/console/0.log" Dec 10 19:12:57 crc kubenswrapper[4828]: I1210 19:12:57.896299 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:12:57 crc kubenswrapper[4828]: I1210 19:12:57.906875 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-serving-cert\") pod \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " Dec 10 19:12:57 crc kubenswrapper[4828]: I1210 19:12:57.906929 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-service-ca\") pod \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " Dec 10 19:12:57 crc kubenswrapper[4828]: I1210 19:12:57.906961 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-oauth-serving-cert\") pod \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " Dec 10 19:12:57 crc kubenswrapper[4828]: I1210 19:12:57.907011 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpmfk\" (UniqueName: \"kubernetes.io/projected/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-kube-api-access-hpmfk\") pod \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " Dec 10 19:12:57 crc kubenswrapper[4828]: I1210 19:12:57.907048 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-config\") pod \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " Dec 10 19:12:57 crc kubenswrapper[4828]: I1210 19:12:57.907073 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-trusted-ca-bundle\") pod \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " Dec 10 19:12:57 crc kubenswrapper[4828]: I1210 19:12:57.907090 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-oauth-config\") pod \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\" (UID: \"434cf9e4-da1b-49cb-a641-9a84ebe2e22c\") " Dec 10 19:12:57 crc kubenswrapper[4828]: I1210 19:12:57.908305 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-service-ca" (OuterVolumeSpecName: "service-ca") pod "434cf9e4-da1b-49cb-a641-9a84ebe2e22c" (UID: "434cf9e4-da1b-49cb-a641-9a84ebe2e22c"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:12:57 crc kubenswrapper[4828]: I1210 19:12:57.908370 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-config" (OuterVolumeSpecName: "console-config") pod "434cf9e4-da1b-49cb-a641-9a84ebe2e22c" (UID: "434cf9e4-da1b-49cb-a641-9a84ebe2e22c"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:12:57 crc kubenswrapper[4828]: I1210 19:12:57.908484 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "434cf9e4-da1b-49cb-a641-9a84ebe2e22c" (UID: "434cf9e4-da1b-49cb-a641-9a84ebe2e22c"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:12:57 crc kubenswrapper[4828]: I1210 19:12:57.908760 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "434cf9e4-da1b-49cb-a641-9a84ebe2e22c" (UID: "434cf9e4-da1b-49cb-a641-9a84ebe2e22c"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:12:57 crc kubenswrapper[4828]: I1210 19:12:57.916439 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "434cf9e4-da1b-49cb-a641-9a84ebe2e22c" (UID: "434cf9e4-da1b-49cb-a641-9a84ebe2e22c"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:12:57 crc kubenswrapper[4828]: I1210 19:12:57.916536 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "434cf9e4-da1b-49cb-a641-9a84ebe2e22c" (UID: "434cf9e4-da1b-49cb-a641-9a84ebe2e22c"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:12:57 crc kubenswrapper[4828]: I1210 19:12:57.936257 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-kube-api-access-hpmfk" (OuterVolumeSpecName: "kube-api-access-hpmfk") pod "434cf9e4-da1b-49cb-a641-9a84ebe2e22c" (UID: "434cf9e4-da1b-49cb-a641-9a84ebe2e22c"). InnerVolumeSpecName "kube-api-access-hpmfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:12:58 crc kubenswrapper[4828]: I1210 19:12:58.007926 4828 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:12:58 crc kubenswrapper[4828]: I1210 19:12:58.007960 4828 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:12:58 crc kubenswrapper[4828]: I1210 19:12:58.007970 4828 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:12:58 crc kubenswrapper[4828]: I1210 19:12:58.007979 4828 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 19:12:58 crc kubenswrapper[4828]: I1210 19:12:58.007988 4828 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 19:12:58 crc kubenswrapper[4828]: I1210 19:12:58.007996 4828 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 19:12:58 crc kubenswrapper[4828]: I1210 19:12:58.008007 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpmfk\" (UniqueName: \"kubernetes.io/projected/434cf9e4-da1b-49cb-a641-9a84ebe2e22c-kube-api-access-hpmfk\") on node \"crc\" DevicePath \"\"" Dec 10 19:12:58 crc kubenswrapper[4828]: I1210 19:12:58.433033 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-8486f475fd-6jrxr_434cf9e4-da1b-49cb-a641-9a84ebe2e22c/console/0.log" Dec 10 19:12:58 crc kubenswrapper[4828]: I1210 19:12:58.433263 4828 generic.go:334] "Generic (PLEG): container finished" podID="434cf9e4-da1b-49cb-a641-9a84ebe2e22c" containerID="2e35b4ad0b8cb77ea24d1d1476570efc4ac30280c569dc13dfaf10b670724d9b" exitCode=2 Dec 10 19:12:58 crc kubenswrapper[4828]: I1210 19:12:58.433296 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8486f475fd-6jrxr" event={"ID":"434cf9e4-da1b-49cb-a641-9a84ebe2e22c","Type":"ContainerDied","Data":"2e35b4ad0b8cb77ea24d1d1476570efc4ac30280c569dc13dfaf10b670724d9b"} Dec 10 19:12:58 crc kubenswrapper[4828]: I1210 19:12:58.433322 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8486f475fd-6jrxr" event={"ID":"434cf9e4-da1b-49cb-a641-9a84ebe2e22c","Type":"ContainerDied","Data":"04f8d8d109b5226c8e13e3b971167463ebca67ee078e1c428abe21f67f85b568"} Dec 10 19:12:58 crc kubenswrapper[4828]: I1210 19:12:58.433334 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8486f475fd-6jrxr" Dec 10 19:12:58 crc kubenswrapper[4828]: I1210 19:12:58.433343 4828 scope.go:117] "RemoveContainer" containerID="2e35b4ad0b8cb77ea24d1d1476570efc4ac30280c569dc13dfaf10b670724d9b" Dec 10 19:12:58 crc kubenswrapper[4828]: I1210 19:12:58.494497 4828 scope.go:117] "RemoveContainer" containerID="2e35b4ad0b8cb77ea24d1d1476570efc4ac30280c569dc13dfaf10b670724d9b" Dec 10 19:12:58 crc kubenswrapper[4828]: E1210 19:12:58.496255 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e35b4ad0b8cb77ea24d1d1476570efc4ac30280c569dc13dfaf10b670724d9b\": container with ID starting with 2e35b4ad0b8cb77ea24d1d1476570efc4ac30280c569dc13dfaf10b670724d9b not found: ID does not exist" containerID="2e35b4ad0b8cb77ea24d1d1476570efc4ac30280c569dc13dfaf10b670724d9b" Dec 10 19:12:58 crc kubenswrapper[4828]: I1210 19:12:58.496309 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e35b4ad0b8cb77ea24d1d1476570efc4ac30280c569dc13dfaf10b670724d9b"} err="failed to get container status \"2e35b4ad0b8cb77ea24d1d1476570efc4ac30280c569dc13dfaf10b670724d9b\": rpc error: code = NotFound desc = could not find container \"2e35b4ad0b8cb77ea24d1d1476570efc4ac30280c569dc13dfaf10b670724d9b\": container with ID starting with 2e35b4ad0b8cb77ea24d1d1476570efc4ac30280c569dc13dfaf10b670724d9b not found: ID does not exist" Dec 10 19:12:58 crc kubenswrapper[4828]: I1210 19:12:58.498715 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-8486f475fd-6jrxr"] Dec 10 19:12:58 crc kubenswrapper[4828]: I1210 19:12:58.511822 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-8486f475fd-6jrxr"] Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.146567 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9"] Dec 10 19:12:59 crc kubenswrapper[4828]: E1210 19:12:59.146936 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="434cf9e4-da1b-49cb-a641-9a84ebe2e22c" containerName="console" Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.146959 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="434cf9e4-da1b-49cb-a641-9a84ebe2e22c" containerName="console" Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.147125 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="434cf9e4-da1b-49cb-a641-9a84ebe2e22c" containerName="console" Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.148337 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.149778 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.158758 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9"] Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.232069 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9\" (UID: \"1f7d42e8-5aae-40e3-a5f8-bd398b69db43\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.232119 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n85jw\" (UniqueName: \"kubernetes.io/projected/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-kube-api-access-n85jw\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9\" (UID: \"1f7d42e8-5aae-40e3-a5f8-bd398b69db43\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.232188 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9\" (UID: \"1f7d42e8-5aae-40e3-a5f8-bd398b69db43\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.333659 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9\" (UID: \"1f7d42e8-5aae-40e3-a5f8-bd398b69db43\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.333760 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9\" (UID: \"1f7d42e8-5aae-40e3-a5f8-bd398b69db43\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.333783 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n85jw\" (UniqueName: \"kubernetes.io/projected/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-kube-api-access-n85jw\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9\" (UID: \"1f7d42e8-5aae-40e3-a5f8-bd398b69db43\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.334203 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9\" (UID: \"1f7d42e8-5aae-40e3-a5f8-bd398b69db43\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.334202 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9\" (UID: \"1f7d42e8-5aae-40e3-a5f8-bd398b69db43\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.355361 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n85jw\" (UniqueName: \"kubernetes.io/projected/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-kube-api-access-n85jw\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9\" (UID: \"1f7d42e8-5aae-40e3-a5f8-bd398b69db43\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.479195 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.798491 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="434cf9e4-da1b-49cb-a641-9a84ebe2e22c" path="/var/lib/kubelet/pods/434cf9e4-da1b-49cb-a641-9a84ebe2e22c/volumes" Dec 10 19:12:59 crc kubenswrapper[4828]: I1210 19:12:59.977928 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9"] Dec 10 19:13:00 crc kubenswrapper[4828]: I1210 19:13:00.451644 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" event={"ID":"1f7d42e8-5aae-40e3-a5f8-bd398b69db43","Type":"ContainerStarted","Data":"620bfde0ffcaac7f2d05a3c6726e2ed67b4f75f26c344219623951a73711d412"} Dec 10 19:13:00 crc kubenswrapper[4828]: I1210 19:13:00.451975 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" event={"ID":"1f7d42e8-5aae-40e3-a5f8-bd398b69db43","Type":"ContainerStarted","Data":"b0912f0f239dfadd0f898d08436ce4f51eba9b7f879cc380e6955368f9f1efae"} Dec 10 19:13:01 crc kubenswrapper[4828]: I1210 19:13:01.462009 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f7d42e8-5aae-40e3-a5f8-bd398b69db43" containerID="620bfde0ffcaac7f2d05a3c6726e2ed67b4f75f26c344219623951a73711d412" exitCode=0 Dec 10 19:13:01 crc kubenswrapper[4828]: I1210 19:13:01.462061 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" event={"ID":"1f7d42e8-5aae-40e3-a5f8-bd398b69db43","Type":"ContainerDied","Data":"620bfde0ffcaac7f2d05a3c6726e2ed67b4f75f26c344219623951a73711d412"} Dec 10 19:13:05 crc kubenswrapper[4828]: I1210 19:13:05.490813 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f7d42e8-5aae-40e3-a5f8-bd398b69db43" containerID="acea7a5e7051bb632e081ef6ac8ceb1cd05b572f1c76e6c2dfb176922d0567bb" exitCode=0 Dec 10 19:13:05 crc kubenswrapper[4828]: I1210 19:13:05.491019 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" event={"ID":"1f7d42e8-5aae-40e3-a5f8-bd398b69db43","Type":"ContainerDied","Data":"acea7a5e7051bb632e081ef6ac8ceb1cd05b572f1c76e6c2dfb176922d0567bb"} Dec 10 19:13:08 crc kubenswrapper[4828]: I1210 19:13:08.512635 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f7d42e8-5aae-40e3-a5f8-bd398b69db43" containerID="49812a46de8a3dd7947f6891834f679a26db5b46f68c036fe8d6586fa2c27a6a" exitCode=0 Dec 10 19:13:08 crc kubenswrapper[4828]: I1210 19:13:08.512741 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" event={"ID":"1f7d42e8-5aae-40e3-a5f8-bd398b69db43","Type":"ContainerDied","Data":"49812a46de8a3dd7947f6891834f679a26db5b46f68c036fe8d6586fa2c27a6a"} Dec 10 19:13:09 crc kubenswrapper[4828]: I1210 19:13:09.817969 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" Dec 10 19:13:09 crc kubenswrapper[4828]: I1210 19:13:09.995754 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n85jw\" (UniqueName: \"kubernetes.io/projected/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-kube-api-access-n85jw\") pod \"1f7d42e8-5aae-40e3-a5f8-bd398b69db43\" (UID: \"1f7d42e8-5aae-40e3-a5f8-bd398b69db43\") " Dec 10 19:13:09 crc kubenswrapper[4828]: I1210 19:13:09.995891 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-bundle\") pod \"1f7d42e8-5aae-40e3-a5f8-bd398b69db43\" (UID: \"1f7d42e8-5aae-40e3-a5f8-bd398b69db43\") " Dec 10 19:13:09 crc kubenswrapper[4828]: I1210 19:13:09.996056 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-util\") pod \"1f7d42e8-5aae-40e3-a5f8-bd398b69db43\" (UID: \"1f7d42e8-5aae-40e3-a5f8-bd398b69db43\") " Dec 10 19:13:09 crc kubenswrapper[4828]: I1210 19:13:09.996771 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-bundle" (OuterVolumeSpecName: "bundle") pod "1f7d42e8-5aae-40e3-a5f8-bd398b69db43" (UID: "1f7d42e8-5aae-40e3-a5f8-bd398b69db43"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:13:10 crc kubenswrapper[4828]: I1210 19:13:10.001293 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-kube-api-access-n85jw" (OuterVolumeSpecName: "kube-api-access-n85jw") pod "1f7d42e8-5aae-40e3-a5f8-bd398b69db43" (UID: "1f7d42e8-5aae-40e3-a5f8-bd398b69db43"). InnerVolumeSpecName "kube-api-access-n85jw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:13:10 crc kubenswrapper[4828]: I1210 19:13:10.007739 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-util" (OuterVolumeSpecName: "util") pod "1f7d42e8-5aae-40e3-a5f8-bd398b69db43" (UID: "1f7d42e8-5aae-40e3-a5f8-bd398b69db43"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:13:10 crc kubenswrapper[4828]: I1210 19:13:10.098191 4828 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-util\") on node \"crc\" DevicePath \"\"" Dec 10 19:13:10 crc kubenswrapper[4828]: I1210 19:13:10.098231 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n85jw\" (UniqueName: \"kubernetes.io/projected/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-kube-api-access-n85jw\") on node \"crc\" DevicePath \"\"" Dec 10 19:13:10 crc kubenswrapper[4828]: I1210 19:13:10.098245 4828 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1f7d42e8-5aae-40e3-a5f8-bd398b69db43-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:13:10 crc kubenswrapper[4828]: I1210 19:13:10.535075 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" event={"ID":"1f7d42e8-5aae-40e3-a5f8-bd398b69db43","Type":"ContainerDied","Data":"b0912f0f239dfadd0f898d08436ce4f51eba9b7f879cc380e6955368f9f1efae"} Dec 10 19:13:10 crc kubenswrapper[4828]: I1210 19:13:10.535478 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0912f0f239dfadd0f898d08436ce4f51eba9b7f879cc380e6955368f9f1efae" Dec 10 19:13:10 crc kubenswrapper[4828]: I1210 19:13:10.535596 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9" Dec 10 19:13:10 crc kubenswrapper[4828]: E1210 19:13:10.646460 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f7d42e8_5aae_40e3_a5f8_bd398b69db43.slice/crio-b0912f0f239dfadd0f898d08436ce4f51eba9b7f879cc380e6955368f9f1efae\": RecentStats: unable to find data in memory cache]" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.512218 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw"] Dec 10 19:13:22 crc kubenswrapper[4828]: E1210 19:13:22.513145 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f7d42e8-5aae-40e3-a5f8-bd398b69db43" containerName="extract" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.513161 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f7d42e8-5aae-40e3-a5f8-bd398b69db43" containerName="extract" Dec 10 19:13:22 crc kubenswrapper[4828]: E1210 19:13:22.513185 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f7d42e8-5aae-40e3-a5f8-bd398b69db43" containerName="pull" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.513191 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f7d42e8-5aae-40e3-a5f8-bd398b69db43" containerName="pull" Dec 10 19:13:22 crc kubenswrapper[4828]: E1210 19:13:22.513202 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f7d42e8-5aae-40e3-a5f8-bd398b69db43" containerName="util" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.513213 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f7d42e8-5aae-40e3-a5f8-bd398b69db43" containerName="util" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.513348 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f7d42e8-5aae-40e3-a5f8-bd398b69db43" containerName="extract" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.513859 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.515668 4828 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.515718 4828 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.516337 4828 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-zcdzx" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.516424 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.516966 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.526280 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw"] Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.686569 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4a495b9b-2ebd-43bd-bb0c-64f32ad257ae-apiservice-cert\") pod \"metallb-operator-controller-manager-75b57b9fc5-tdblw\" (UID: \"4a495b9b-2ebd-43bd-bb0c-64f32ad257ae\") " pod="metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.686627 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4a495b9b-2ebd-43bd-bb0c-64f32ad257ae-webhook-cert\") pod \"metallb-operator-controller-manager-75b57b9fc5-tdblw\" (UID: \"4a495b9b-2ebd-43bd-bb0c-64f32ad257ae\") " pod="metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.686673 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knssr\" (UniqueName: \"kubernetes.io/projected/4a495b9b-2ebd-43bd-bb0c-64f32ad257ae-kube-api-access-knssr\") pod \"metallb-operator-controller-manager-75b57b9fc5-tdblw\" (UID: \"4a495b9b-2ebd-43bd-bb0c-64f32ad257ae\") " pod="metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.753881 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8"] Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.755012 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.758740 4828 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.759121 4828 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.759190 4828 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-vz79q" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.773224 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8"] Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.788658 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knssr\" (UniqueName: \"kubernetes.io/projected/4a495b9b-2ebd-43bd-bb0c-64f32ad257ae-kube-api-access-knssr\") pod \"metallb-operator-controller-manager-75b57b9fc5-tdblw\" (UID: \"4a495b9b-2ebd-43bd-bb0c-64f32ad257ae\") " pod="metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.788771 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4a495b9b-2ebd-43bd-bb0c-64f32ad257ae-apiservice-cert\") pod \"metallb-operator-controller-manager-75b57b9fc5-tdblw\" (UID: \"4a495b9b-2ebd-43bd-bb0c-64f32ad257ae\") " pod="metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.788808 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4a495b9b-2ebd-43bd-bb0c-64f32ad257ae-webhook-cert\") pod \"metallb-operator-controller-manager-75b57b9fc5-tdblw\" (UID: \"4a495b9b-2ebd-43bd-bb0c-64f32ad257ae\") " pod="metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.795327 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4a495b9b-2ebd-43bd-bb0c-64f32ad257ae-webhook-cert\") pod \"metallb-operator-controller-manager-75b57b9fc5-tdblw\" (UID: \"4a495b9b-2ebd-43bd-bb0c-64f32ad257ae\") " pod="metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.806225 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4a495b9b-2ebd-43bd-bb0c-64f32ad257ae-apiservice-cert\") pod \"metallb-operator-controller-manager-75b57b9fc5-tdblw\" (UID: \"4a495b9b-2ebd-43bd-bb0c-64f32ad257ae\") " pod="metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.810865 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knssr\" (UniqueName: \"kubernetes.io/projected/4a495b9b-2ebd-43bd-bb0c-64f32ad257ae-kube-api-access-knssr\") pod \"metallb-operator-controller-manager-75b57b9fc5-tdblw\" (UID: \"4a495b9b-2ebd-43bd-bb0c-64f32ad257ae\") " pod="metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.830604 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.889918 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/87eba563-d95b-486f-bde7-61d7a764d996-webhook-cert\") pod \"metallb-operator-webhook-server-5bd8d89b5d-tdlk8\" (UID: \"87eba563-d95b-486f-bde7-61d7a764d996\") " pod="metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.890055 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4bkf\" (UniqueName: \"kubernetes.io/projected/87eba563-d95b-486f-bde7-61d7a764d996-kube-api-access-c4bkf\") pod \"metallb-operator-webhook-server-5bd8d89b5d-tdlk8\" (UID: \"87eba563-d95b-486f-bde7-61d7a764d996\") " pod="metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.890148 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/87eba563-d95b-486f-bde7-61d7a764d996-apiservice-cert\") pod \"metallb-operator-webhook-server-5bd8d89b5d-tdlk8\" (UID: \"87eba563-d95b-486f-bde7-61d7a764d996\") " pod="metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.991962 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/87eba563-d95b-486f-bde7-61d7a764d996-webhook-cert\") pod \"metallb-operator-webhook-server-5bd8d89b5d-tdlk8\" (UID: \"87eba563-d95b-486f-bde7-61d7a764d996\") " pod="metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.992325 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4bkf\" (UniqueName: \"kubernetes.io/projected/87eba563-d95b-486f-bde7-61d7a764d996-kube-api-access-c4bkf\") pod \"metallb-operator-webhook-server-5bd8d89b5d-tdlk8\" (UID: \"87eba563-d95b-486f-bde7-61d7a764d996\") " pod="metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.992370 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/87eba563-d95b-486f-bde7-61d7a764d996-apiservice-cert\") pod \"metallb-operator-webhook-server-5bd8d89b5d-tdlk8\" (UID: \"87eba563-d95b-486f-bde7-61d7a764d996\") " pod="metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.997254 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/87eba563-d95b-486f-bde7-61d7a764d996-webhook-cert\") pod \"metallb-operator-webhook-server-5bd8d89b5d-tdlk8\" (UID: \"87eba563-d95b-486f-bde7-61d7a764d996\") " pod="metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8" Dec 10 19:13:22 crc kubenswrapper[4828]: I1210 19:13:22.997302 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/87eba563-d95b-486f-bde7-61d7a764d996-apiservice-cert\") pod \"metallb-operator-webhook-server-5bd8d89b5d-tdlk8\" (UID: \"87eba563-d95b-486f-bde7-61d7a764d996\") " pod="metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8" Dec 10 19:13:23 crc kubenswrapper[4828]: I1210 19:13:23.016122 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4bkf\" (UniqueName: \"kubernetes.io/projected/87eba563-d95b-486f-bde7-61d7a764d996-kube-api-access-c4bkf\") pod \"metallb-operator-webhook-server-5bd8d89b5d-tdlk8\" (UID: \"87eba563-d95b-486f-bde7-61d7a764d996\") " pod="metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8" Dec 10 19:13:23 crc kubenswrapper[4828]: I1210 19:13:23.072139 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8" Dec 10 19:13:23 crc kubenswrapper[4828]: I1210 19:13:23.290244 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw"] Dec 10 19:13:23 crc kubenswrapper[4828]: W1210 19:13:23.296929 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a495b9b_2ebd_43bd_bb0c_64f32ad257ae.slice/crio-d99adffa74ae08afcefb6cd00eacdc933d8d793858cbc5275542a49613b3cf0d WatchSource:0}: Error finding container d99adffa74ae08afcefb6cd00eacdc933d8d793858cbc5275542a49613b3cf0d: Status 404 returned error can't find the container with id d99adffa74ae08afcefb6cd00eacdc933d8d793858cbc5275542a49613b3cf0d Dec 10 19:13:23 crc kubenswrapper[4828]: I1210 19:13:23.493018 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8"] Dec 10 19:13:23 crc kubenswrapper[4828]: W1210 19:13:23.494444 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87eba563_d95b_486f_bde7_61d7a764d996.slice/crio-dbd0d25d00a3f941d59b2a63802f656e1c1be9a94964daa084540900dcf4d7c1 WatchSource:0}: Error finding container dbd0d25d00a3f941d59b2a63802f656e1c1be9a94964daa084540900dcf4d7c1: Status 404 returned error can't find the container with id dbd0d25d00a3f941d59b2a63802f656e1c1be9a94964daa084540900dcf4d7c1 Dec 10 19:13:23 crc kubenswrapper[4828]: I1210 19:13:23.619483 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw" event={"ID":"4a495b9b-2ebd-43bd-bb0c-64f32ad257ae","Type":"ContainerStarted","Data":"d99adffa74ae08afcefb6cd00eacdc933d8d793858cbc5275542a49613b3cf0d"} Dec 10 19:13:23 crc kubenswrapper[4828]: I1210 19:13:23.620998 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8" event={"ID":"87eba563-d95b-486f-bde7-61d7a764d996","Type":"ContainerStarted","Data":"dbd0d25d00a3f941d59b2a63802f656e1c1be9a94964daa084540900dcf4d7c1"} Dec 10 19:13:28 crc kubenswrapper[4828]: I1210 19:13:28.666550 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw" event={"ID":"4a495b9b-2ebd-43bd-bb0c-64f32ad257ae","Type":"ContainerStarted","Data":"5bbc3dd821af8d353c022b3d58686b849916c03c69be5023e6bb24d100df4bd4"} Dec 10 19:13:28 crc kubenswrapper[4828]: I1210 19:13:28.667229 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw" Dec 10 19:13:28 crc kubenswrapper[4828]: I1210 19:13:28.696830 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw" podStartSLOduration=1.6627075489999998 podStartE2EDuration="6.696665682s" podCreationTimestamp="2025-12-10 19:13:22 +0000 UTC" firstStartedPulling="2025-12-10 19:13:23.300420362 +0000 UTC m=+1083.811031367" lastFinishedPulling="2025-12-10 19:13:28.334378495 +0000 UTC m=+1088.844989500" observedRunningTime="2025-12-10 19:13:28.688997572 +0000 UTC m=+1089.199608597" watchObservedRunningTime="2025-12-10 19:13:28.696665682 +0000 UTC m=+1089.207276687" Dec 10 19:13:29 crc kubenswrapper[4828]: I1210 19:13:29.676047 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8" event={"ID":"87eba563-d95b-486f-bde7-61d7a764d996","Type":"ContainerStarted","Data":"f67cdbbe287cf875232963c846584db3ff4c8fb742a835136e59672f293d1ae7"} Dec 10 19:13:29 crc kubenswrapper[4828]: I1210 19:13:29.676366 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8" Dec 10 19:13:43 crc kubenswrapper[4828]: I1210 19:13:43.076499 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8" Dec 10 19:13:43 crc kubenswrapper[4828]: I1210 19:13:43.111988 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8" podStartSLOduration=16.257469973 podStartE2EDuration="21.111973427s" podCreationTimestamp="2025-12-10 19:13:22 +0000 UTC" firstStartedPulling="2025-12-10 19:13:23.497326866 +0000 UTC m=+1084.007937871" lastFinishedPulling="2025-12-10 19:13:28.35183032 +0000 UTC m=+1088.862441325" observedRunningTime="2025-12-10 19:13:29.696310548 +0000 UTC m=+1090.206921553" watchObservedRunningTime="2025-12-10 19:13:43.111973427 +0000 UTC m=+1103.622584432" Dec 10 19:13:51 crc kubenswrapper[4828]: I1210 19:13:51.230465 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:13:51 crc kubenswrapper[4828]: I1210 19:13:51.231046 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:14:02 crc kubenswrapper[4828]: I1210 19:14:02.843361 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-75b57b9fc5-tdblw" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.541326 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-97pxg"] Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.546430 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.548575 4828 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.548790 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.549235 4828 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-nxl2s" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.560856 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-hzjlj"] Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.562056 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hzjlj" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.564777 4828 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.588069 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-hzjlj"] Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.645825 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-wjndz"] Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.647336 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-wjndz" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.648888 4828 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.649050 4828 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.649313 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.649438 4828 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-qtd95" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.650321 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-frr-sockets\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.650349 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-frr-startup\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.650422 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-metrics\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.650443 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-metrics-certs\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.650485 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-frr-conf\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.650502 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-reloader\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.650543 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp6f9\" (UniqueName: \"kubernetes.io/projected/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-kube-api-access-vp6f9\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.672383 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-h4cp8"] Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.673518 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-h4cp8" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.677526 4828 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.709713 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-h4cp8"] Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.752312 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-frr-conf\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.752364 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-reloader\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.752401 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-metrics-certs\") pod \"speaker-wjndz\" (UID: \"dd16c5ff-d9b5-4df0-a3ca-b7900966bc37\") " pod="metallb-system/speaker-wjndz" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.752431 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf9xw\" (UniqueName: \"kubernetes.io/projected/159d6965-2b9c-461f-8f20-cf338919fcd7-kube-api-access-zf9xw\") pod \"frr-k8s-webhook-server-7fcb986d4-hzjlj\" (UID: \"159d6965-2b9c-461f-8f20-cf338919fcd7\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hzjlj" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.752490 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp6f9\" (UniqueName: \"kubernetes.io/projected/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-kube-api-access-vp6f9\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.752527 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-frr-sockets\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.752545 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-frr-startup\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.752569 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/159d6965-2b9c-461f-8f20-cf338919fcd7-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-hzjlj\" (UID: \"159d6965-2b9c-461f-8f20-cf338919fcd7\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hzjlj" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.752619 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-metallb-excludel2\") pod \"speaker-wjndz\" (UID: \"dd16c5ff-d9b5-4df0-a3ca-b7900966bc37\") " pod="metallb-system/speaker-wjndz" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.752640 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-memberlist\") pod \"speaker-wjndz\" (UID: \"dd16c5ff-d9b5-4df0-a3ca-b7900966bc37\") " pod="metallb-system/speaker-wjndz" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.752689 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-metrics\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.752710 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctblp\" (UniqueName: \"kubernetes.io/projected/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-kube-api-access-ctblp\") pod \"speaker-wjndz\" (UID: \"dd16c5ff-d9b5-4df0-a3ca-b7900966bc37\") " pod="metallb-system/speaker-wjndz" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.752736 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-metrics-certs\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.752757 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-frr-conf\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.753033 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-reloader\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: E1210 19:14:03.753146 4828 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 10 19:14:03 crc kubenswrapper[4828]: E1210 19:14:03.753202 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-metrics-certs podName:74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab nodeName:}" failed. No retries permitted until 2025-12-10 19:14:04.253185326 +0000 UTC m=+1124.763796331 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-metrics-certs") pod "frr-k8s-97pxg" (UID: "74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab") : secret "frr-k8s-certs-secret" not found Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.753276 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-metrics\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.753707 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-frr-sockets\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.754019 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-frr-startup\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.776154 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp6f9\" (UniqueName: \"kubernetes.io/projected/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-kube-api-access-vp6f9\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.854602 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/159d6965-2b9c-461f-8f20-cf338919fcd7-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-hzjlj\" (UID: \"159d6965-2b9c-461f-8f20-cf338919fcd7\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hzjlj" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.854661 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-metallb-excludel2\") pod \"speaker-wjndz\" (UID: \"dd16c5ff-d9b5-4df0-a3ca-b7900966bc37\") " pod="metallb-system/speaker-wjndz" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.854681 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-memberlist\") pod \"speaker-wjndz\" (UID: \"dd16c5ff-d9b5-4df0-a3ca-b7900966bc37\") " pod="metallb-system/speaker-wjndz" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.854723 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctblp\" (UniqueName: \"kubernetes.io/projected/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-kube-api-access-ctblp\") pod \"speaker-wjndz\" (UID: \"dd16c5ff-d9b5-4df0-a3ca-b7900966bc37\") " pod="metallb-system/speaker-wjndz" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.854785 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-metrics-certs\") pod \"speaker-wjndz\" (UID: \"dd16c5ff-d9b5-4df0-a3ca-b7900966bc37\") " pod="metallb-system/speaker-wjndz" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.854820 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf9xw\" (UniqueName: \"kubernetes.io/projected/159d6965-2b9c-461f-8f20-cf338919fcd7-kube-api-access-zf9xw\") pod \"frr-k8s-webhook-server-7fcb986d4-hzjlj\" (UID: \"159d6965-2b9c-461f-8f20-cf338919fcd7\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hzjlj" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.854850 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdn9n\" (UniqueName: \"kubernetes.io/projected/0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c-kube-api-access-tdn9n\") pod \"controller-f8648f98b-h4cp8\" (UID: \"0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c\") " pod="metallb-system/controller-f8648f98b-h4cp8" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.854872 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c-cert\") pod \"controller-f8648f98b-h4cp8\" (UID: \"0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c\") " pod="metallb-system/controller-f8648f98b-h4cp8" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.854896 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c-metrics-certs\") pod \"controller-f8648f98b-h4cp8\" (UID: \"0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c\") " pod="metallb-system/controller-f8648f98b-h4cp8" Dec 10 19:14:03 crc kubenswrapper[4828]: E1210 19:14:03.855056 4828 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 10 19:14:03 crc kubenswrapper[4828]: E1210 19:14:03.855097 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-memberlist podName:dd16c5ff-d9b5-4df0-a3ca-b7900966bc37 nodeName:}" failed. No retries permitted until 2025-12-10 19:14:04.355082982 +0000 UTC m=+1124.865693987 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-memberlist") pod "speaker-wjndz" (UID: "dd16c5ff-d9b5-4df0-a3ca-b7900966bc37") : secret "metallb-memberlist" not found Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.855480 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-metallb-excludel2\") pod \"speaker-wjndz\" (UID: \"dd16c5ff-d9b5-4df0-a3ca-b7900966bc37\") " pod="metallb-system/speaker-wjndz" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.858296 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/159d6965-2b9c-461f-8f20-cf338919fcd7-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-hzjlj\" (UID: \"159d6965-2b9c-461f-8f20-cf338919fcd7\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hzjlj" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.858890 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-metrics-certs\") pod \"speaker-wjndz\" (UID: \"dd16c5ff-d9b5-4df0-a3ca-b7900966bc37\") " pod="metallb-system/speaker-wjndz" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.877023 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctblp\" (UniqueName: \"kubernetes.io/projected/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-kube-api-access-ctblp\") pod \"speaker-wjndz\" (UID: \"dd16c5ff-d9b5-4df0-a3ca-b7900966bc37\") " pod="metallb-system/speaker-wjndz" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.877367 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf9xw\" (UniqueName: \"kubernetes.io/projected/159d6965-2b9c-461f-8f20-cf338919fcd7-kube-api-access-zf9xw\") pod \"frr-k8s-webhook-server-7fcb986d4-hzjlj\" (UID: \"159d6965-2b9c-461f-8f20-cf338919fcd7\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hzjlj" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.881043 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hzjlj" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.956120 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdn9n\" (UniqueName: \"kubernetes.io/projected/0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c-kube-api-access-tdn9n\") pod \"controller-f8648f98b-h4cp8\" (UID: \"0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c\") " pod="metallb-system/controller-f8648f98b-h4cp8" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.956359 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c-cert\") pod \"controller-f8648f98b-h4cp8\" (UID: \"0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c\") " pod="metallb-system/controller-f8648f98b-h4cp8" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.956450 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c-metrics-certs\") pod \"controller-f8648f98b-h4cp8\" (UID: \"0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c\") " pod="metallb-system/controller-f8648f98b-h4cp8" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.961459 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c-cert\") pod \"controller-f8648f98b-h4cp8\" (UID: \"0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c\") " pod="metallb-system/controller-f8648f98b-h4cp8" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.972213 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c-metrics-certs\") pod \"controller-f8648f98b-h4cp8\" (UID: \"0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c\") " pod="metallb-system/controller-f8648f98b-h4cp8" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.975725 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdn9n\" (UniqueName: \"kubernetes.io/projected/0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c-kube-api-access-tdn9n\") pod \"controller-f8648f98b-h4cp8\" (UID: \"0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c\") " pod="metallb-system/controller-f8648f98b-h4cp8" Dec 10 19:14:03 crc kubenswrapper[4828]: I1210 19:14:03.989558 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-h4cp8" Dec 10 19:14:04 crc kubenswrapper[4828]: I1210 19:14:04.262735 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-metrics-certs\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:04 crc kubenswrapper[4828]: I1210 19:14:04.267331 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab-metrics-certs\") pod \"frr-k8s-97pxg\" (UID: \"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab\") " pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:04 crc kubenswrapper[4828]: I1210 19:14:04.294502 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-hzjlj"] Dec 10 19:14:04 crc kubenswrapper[4828]: I1210 19:14:04.364596 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-memberlist\") pod \"speaker-wjndz\" (UID: \"dd16c5ff-d9b5-4df0-a3ca-b7900966bc37\") " pod="metallb-system/speaker-wjndz" Dec 10 19:14:04 crc kubenswrapper[4828]: E1210 19:14:04.364840 4828 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 10 19:14:04 crc kubenswrapper[4828]: E1210 19:14:04.365123 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-memberlist podName:dd16c5ff-d9b5-4df0-a3ca-b7900966bc37 nodeName:}" failed. No retries permitted until 2025-12-10 19:14:05.365101181 +0000 UTC m=+1125.875712186 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-memberlist") pod "speaker-wjndz" (UID: "dd16c5ff-d9b5-4df0-a3ca-b7900966bc37") : secret "metallb-memberlist" not found Dec 10 19:14:04 crc kubenswrapper[4828]: I1210 19:14:04.393058 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-h4cp8"] Dec 10 19:14:04 crc kubenswrapper[4828]: W1210 19:14:04.395943 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0aeb78a5_f8f6_4cdc_ba8f_32d5fc763d6c.slice/crio-a42084b00026d5fe634c40312066afa1384a90d83e9914d73b8da75f6abde5e9 WatchSource:0}: Error finding container a42084b00026d5fe634c40312066afa1384a90d83e9914d73b8da75f6abde5e9: Status 404 returned error can't find the container with id a42084b00026d5fe634c40312066afa1384a90d83e9914d73b8da75f6abde5e9 Dec 10 19:14:04 crc kubenswrapper[4828]: I1210 19:14:04.466393 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:04 crc kubenswrapper[4828]: I1210 19:14:04.926700 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-h4cp8" event={"ID":"0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c","Type":"ContainerStarted","Data":"a42084b00026d5fe634c40312066afa1384a90d83e9914d73b8da75f6abde5e9"} Dec 10 19:14:04 crc kubenswrapper[4828]: I1210 19:14:04.928676 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hzjlj" event={"ID":"159d6965-2b9c-461f-8f20-cf338919fcd7","Type":"ContainerStarted","Data":"b0195b91dae31914a28d9df75eeb0d5200f407b7b61fef8782fc6b996ca40a38"} Dec 10 19:14:05 crc kubenswrapper[4828]: I1210 19:14:05.380259 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-memberlist\") pod \"speaker-wjndz\" (UID: \"dd16c5ff-d9b5-4df0-a3ca-b7900966bc37\") " pod="metallb-system/speaker-wjndz" Dec 10 19:14:05 crc kubenswrapper[4828]: I1210 19:14:05.386133 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/dd16c5ff-d9b5-4df0-a3ca-b7900966bc37-memberlist\") pod \"speaker-wjndz\" (UID: \"dd16c5ff-d9b5-4df0-a3ca-b7900966bc37\") " pod="metallb-system/speaker-wjndz" Dec 10 19:14:05 crc kubenswrapper[4828]: I1210 19:14:05.467755 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-wjndz" Dec 10 19:14:05 crc kubenswrapper[4828]: W1210 19:14:05.497313 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd16c5ff_d9b5_4df0_a3ca_b7900966bc37.slice/crio-76603a4e2395a1af37496404e5351278c4b268cc26046999cdc7d3c13d614d1e WatchSource:0}: Error finding container 76603a4e2395a1af37496404e5351278c4b268cc26046999cdc7d3c13d614d1e: Status 404 returned error can't find the container with id 76603a4e2395a1af37496404e5351278c4b268cc26046999cdc7d3c13d614d1e Dec 10 19:14:05 crc kubenswrapper[4828]: I1210 19:14:05.942742 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-h4cp8" event={"ID":"0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c","Type":"ContainerStarted","Data":"3893b2e1ec2e133cfc850cd1213e83a5605f7e7bab3628a5cf731a47a35bcd50"} Dec 10 19:14:05 crc kubenswrapper[4828]: I1210 19:14:05.942830 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-h4cp8" event={"ID":"0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c","Type":"ContainerStarted","Data":"d8b40b2da145effe159a3f2482dd977230121b2e44bbf9066cac2c6b936d58ca"} Dec 10 19:14:05 crc kubenswrapper[4828]: I1210 19:14:05.942876 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-h4cp8" Dec 10 19:14:05 crc kubenswrapper[4828]: I1210 19:14:05.944558 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-97pxg" event={"ID":"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab","Type":"ContainerStarted","Data":"987a6438f7efa4b2c5d6c58b42606f02446946208b4a5a5c0230b7b306d4244e"} Dec 10 19:14:05 crc kubenswrapper[4828]: I1210 19:14:05.945635 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wjndz" event={"ID":"dd16c5ff-d9b5-4df0-a3ca-b7900966bc37","Type":"ContainerStarted","Data":"76603a4e2395a1af37496404e5351278c4b268cc26046999cdc7d3c13d614d1e"} Dec 10 19:14:05 crc kubenswrapper[4828]: I1210 19:14:05.961114 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-h4cp8" podStartSLOduration=2.961099598 podStartE2EDuration="2.961099598s" podCreationTimestamp="2025-12-10 19:14:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:14:05.959930758 +0000 UTC m=+1126.470541763" watchObservedRunningTime="2025-12-10 19:14:05.961099598 +0000 UTC m=+1126.471710603" Dec 10 19:14:06 crc kubenswrapper[4828]: I1210 19:14:06.956021 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wjndz" event={"ID":"dd16c5ff-d9b5-4df0-a3ca-b7900966bc37","Type":"ContainerStarted","Data":"d2efe7220f800e73641492a0f906b29538fee17b6ea141bf9256f97ac29fe08f"} Dec 10 19:14:06 crc kubenswrapper[4828]: I1210 19:14:06.956301 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-wjndz" event={"ID":"dd16c5ff-d9b5-4df0-a3ca-b7900966bc37","Type":"ContainerStarted","Data":"d3c285e50082441569eb306997612ef5390b6ca3fc55ff0ddfd00736d10fb204"} Dec 10 19:14:06 crc kubenswrapper[4828]: I1210 19:14:06.956319 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-wjndz" Dec 10 19:14:06 crc kubenswrapper[4828]: I1210 19:14:06.981981 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-wjndz" podStartSLOduration=3.981962328 podStartE2EDuration="3.981962328s" podCreationTimestamp="2025-12-10 19:14:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:14:06.977485822 +0000 UTC m=+1127.488096847" watchObservedRunningTime="2025-12-10 19:14:06.981962328 +0000 UTC m=+1127.492573333" Dec 10 19:14:11 crc kubenswrapper[4828]: I1210 19:14:11.992490 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hzjlj" event={"ID":"159d6965-2b9c-461f-8f20-cf338919fcd7","Type":"ContainerStarted","Data":"60ef404fee8df36dcde24f340c3908b8a416b1c4336eaa742a7fdeda24b26729"} Dec 10 19:14:11 crc kubenswrapper[4828]: I1210 19:14:11.993112 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hzjlj" Dec 10 19:14:11 crc kubenswrapper[4828]: I1210 19:14:11.993987 4828 generic.go:334] "Generic (PLEG): container finished" podID="74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab" containerID="75905b59c97af199b2383233952ccbce5b3a8816c4a3c4263fecd6d455162c95" exitCode=0 Dec 10 19:14:11 crc kubenswrapper[4828]: I1210 19:14:11.994017 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-97pxg" event={"ID":"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab","Type":"ContainerDied","Data":"75905b59c97af199b2383233952ccbce5b3a8816c4a3c4263fecd6d455162c95"} Dec 10 19:14:12 crc kubenswrapper[4828]: I1210 19:14:12.008974 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hzjlj" podStartSLOduration=2.328246456 podStartE2EDuration="9.008958889s" podCreationTimestamp="2025-12-10 19:14:03 +0000 UTC" firstStartedPulling="2025-12-10 19:14:04.29716471 +0000 UTC m=+1124.807775715" lastFinishedPulling="2025-12-10 19:14:10.977877143 +0000 UTC m=+1131.488488148" observedRunningTime="2025-12-10 19:14:12.006114735 +0000 UTC m=+1132.516725760" watchObservedRunningTime="2025-12-10 19:14:12.008958889 +0000 UTC m=+1132.519569894" Dec 10 19:14:13 crc kubenswrapper[4828]: I1210 19:14:13.002211 4828 generic.go:334] "Generic (PLEG): container finished" podID="74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab" containerID="196c178b40c03bdddc88b8567e5046f74b71096408f672df15f9f4539f1b915b" exitCode=0 Dec 10 19:14:13 crc kubenswrapper[4828]: I1210 19:14:13.002262 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-97pxg" event={"ID":"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab","Type":"ContainerDied","Data":"196c178b40c03bdddc88b8567e5046f74b71096408f672df15f9f4539f1b915b"} Dec 10 19:14:14 crc kubenswrapper[4828]: I1210 19:14:14.033666 4828 generic.go:334] "Generic (PLEG): container finished" podID="74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab" containerID="7222f55cda3dc1d6173a59b735f506ba826e546f98d6cd56b418d9ed98549034" exitCode=0 Dec 10 19:14:14 crc kubenswrapper[4828]: I1210 19:14:14.034651 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-97pxg" event={"ID":"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab","Type":"ContainerDied","Data":"7222f55cda3dc1d6173a59b735f506ba826e546f98d6cd56b418d9ed98549034"} Dec 10 19:14:15 crc kubenswrapper[4828]: I1210 19:14:15.046840 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-97pxg" event={"ID":"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab","Type":"ContainerStarted","Data":"220f730b7fade6885f16f93b2ccc4fe5b52f81cd85afddaa964d5b02efb15842"} Dec 10 19:14:16 crc kubenswrapper[4828]: I1210 19:14:16.063089 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-97pxg" event={"ID":"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab","Type":"ContainerStarted","Data":"b0b6ea146f3873fefbbc6fad5d77a32211f56f363d7a0dad6a33eb1efc62c472"} Dec 10 19:14:16 crc kubenswrapper[4828]: I1210 19:14:16.063133 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-97pxg" event={"ID":"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab","Type":"ContainerStarted","Data":"9fbaa732a431dd3ad157c6fc8b69078af947fe8db8adb072f10b7ab5e4d1d36b"} Dec 10 19:14:16 crc kubenswrapper[4828]: I1210 19:14:16.063143 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-97pxg" event={"ID":"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab","Type":"ContainerStarted","Data":"dae307f0b74e1ac064edcf54e9c3b4c65ad799a73d33096b62b84c1b3421b8b5"} Dec 10 19:14:16 crc kubenswrapper[4828]: I1210 19:14:16.063151 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-97pxg" event={"ID":"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab","Type":"ContainerStarted","Data":"0c1a7178e26fbe22c1bfacf578d5dbc8f80d9c7172c93f47c037ef6e63e3dc64"} Dec 10 19:14:17 crc kubenswrapper[4828]: I1210 19:14:17.080574 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-97pxg" event={"ID":"74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab","Type":"ContainerStarted","Data":"9e1f58806fb7e76c202779764424d5a9aa75d1323b20d7e3c3cbffd01e46f6c5"} Dec 10 19:14:17 crc kubenswrapper[4828]: I1210 19:14:17.081302 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:17 crc kubenswrapper[4828]: I1210 19:14:17.110514 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-97pxg" podStartSLOduration=8.038494873 podStartE2EDuration="14.110469613s" podCreationTimestamp="2025-12-10 19:14:03 +0000 UTC" firstStartedPulling="2025-12-10 19:14:04.926145201 +0000 UTC m=+1125.436756226" lastFinishedPulling="2025-12-10 19:14:10.998119961 +0000 UTC m=+1131.508730966" observedRunningTime="2025-12-10 19:14:17.103759418 +0000 UTC m=+1137.614370423" watchObservedRunningTime="2025-12-10 19:14:17.110469613 +0000 UTC m=+1137.621080638" Dec 10 19:14:19 crc kubenswrapper[4828]: I1210 19:14:19.467250 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:19 crc kubenswrapper[4828]: I1210 19:14:19.539211 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:21 crc kubenswrapper[4828]: I1210 19:14:21.230315 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:14:21 crc kubenswrapper[4828]: I1210 19:14:21.230665 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:14:23 crc kubenswrapper[4828]: I1210 19:14:23.888241 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-hzjlj" Dec 10 19:14:24 crc kubenswrapper[4828]: I1210 19:14:24.001895 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-h4cp8" Dec 10 19:14:24 crc kubenswrapper[4828]: I1210 19:14:24.471208 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-97pxg" Dec 10 19:14:25 crc kubenswrapper[4828]: I1210 19:14:25.472181 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-wjndz" Dec 10 19:14:28 crc kubenswrapper[4828]: I1210 19:14:28.225113 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-zlxm5"] Dec 10 19:14:28 crc kubenswrapper[4828]: I1210 19:14:28.227615 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zlxm5" Dec 10 19:14:28 crc kubenswrapper[4828]: I1210 19:14:28.231313 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-ntdpw" Dec 10 19:14:28 crc kubenswrapper[4828]: I1210 19:14:28.231628 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 10 19:14:28 crc kubenswrapper[4828]: I1210 19:14:28.236368 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 10 19:14:28 crc kubenswrapper[4828]: I1210 19:14:28.244546 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zlxm5"] Dec 10 19:14:28 crc kubenswrapper[4828]: I1210 19:14:28.287183 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl24j\" (UniqueName: \"kubernetes.io/projected/3f36c8bd-7eea-416b-bf77-5bca884a3426-kube-api-access-dl24j\") pod \"openstack-operator-index-zlxm5\" (UID: \"3f36c8bd-7eea-416b-bf77-5bca884a3426\") " pod="openstack-operators/openstack-operator-index-zlxm5" Dec 10 19:14:28 crc kubenswrapper[4828]: I1210 19:14:28.388423 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl24j\" (UniqueName: \"kubernetes.io/projected/3f36c8bd-7eea-416b-bf77-5bca884a3426-kube-api-access-dl24j\") pod \"openstack-operator-index-zlxm5\" (UID: \"3f36c8bd-7eea-416b-bf77-5bca884a3426\") " pod="openstack-operators/openstack-operator-index-zlxm5" Dec 10 19:14:28 crc kubenswrapper[4828]: I1210 19:14:28.410225 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl24j\" (UniqueName: \"kubernetes.io/projected/3f36c8bd-7eea-416b-bf77-5bca884a3426-kube-api-access-dl24j\") pod \"openstack-operator-index-zlxm5\" (UID: \"3f36c8bd-7eea-416b-bf77-5bca884a3426\") " pod="openstack-operators/openstack-operator-index-zlxm5" Dec 10 19:14:28 crc kubenswrapper[4828]: I1210 19:14:28.551299 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zlxm5" Dec 10 19:14:28 crc kubenswrapper[4828]: I1210 19:14:28.865579 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zlxm5"] Dec 10 19:14:29 crc kubenswrapper[4828]: I1210 19:14:29.184025 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zlxm5" event={"ID":"3f36c8bd-7eea-416b-bf77-5bca884a3426","Type":"ContainerStarted","Data":"8f13fd6253e2da3f22a32fade60ee67b03db9af5203bbd7751d61888a739446f"} Dec 10 19:14:31 crc kubenswrapper[4828]: I1210 19:14:31.578929 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-zlxm5"] Dec 10 19:14:32 crc kubenswrapper[4828]: I1210 19:14:32.184608 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-6m6c8"] Dec 10 19:14:32 crc kubenswrapper[4828]: I1210 19:14:32.185558 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6m6c8" Dec 10 19:14:32 crc kubenswrapper[4828]: I1210 19:14:32.195712 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6m6c8"] Dec 10 19:14:32 crc kubenswrapper[4828]: I1210 19:14:32.301531 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhxjt\" (UniqueName: \"kubernetes.io/projected/eaa2ffd3-6860-4f1f-9ada-04927d8a81eb-kube-api-access-fhxjt\") pod \"openstack-operator-index-6m6c8\" (UID: \"eaa2ffd3-6860-4f1f-9ada-04927d8a81eb\") " pod="openstack-operators/openstack-operator-index-6m6c8" Dec 10 19:14:32 crc kubenswrapper[4828]: I1210 19:14:32.403587 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhxjt\" (UniqueName: \"kubernetes.io/projected/eaa2ffd3-6860-4f1f-9ada-04927d8a81eb-kube-api-access-fhxjt\") pod \"openstack-operator-index-6m6c8\" (UID: \"eaa2ffd3-6860-4f1f-9ada-04927d8a81eb\") " pod="openstack-operators/openstack-operator-index-6m6c8" Dec 10 19:14:32 crc kubenswrapper[4828]: I1210 19:14:32.425852 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhxjt\" (UniqueName: \"kubernetes.io/projected/eaa2ffd3-6860-4f1f-9ada-04927d8a81eb-kube-api-access-fhxjt\") pod \"openstack-operator-index-6m6c8\" (UID: \"eaa2ffd3-6860-4f1f-9ada-04927d8a81eb\") " pod="openstack-operators/openstack-operator-index-6m6c8" Dec 10 19:14:32 crc kubenswrapper[4828]: I1210 19:14:32.510458 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6m6c8" Dec 10 19:14:32 crc kubenswrapper[4828]: I1210 19:14:32.900169 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6m6c8"] Dec 10 19:14:33 crc kubenswrapper[4828]: W1210 19:14:33.132397 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeaa2ffd3_6860_4f1f_9ada_04927d8a81eb.slice/crio-ca155bc7158ca12a9e9559b7e002e882d3e73f90e376f829ce3ffbfb142818fd WatchSource:0}: Error finding container ca155bc7158ca12a9e9559b7e002e882d3e73f90e376f829ce3ffbfb142818fd: Status 404 returned error can't find the container with id ca155bc7158ca12a9e9559b7e002e882d3e73f90e376f829ce3ffbfb142818fd Dec 10 19:14:33 crc kubenswrapper[4828]: I1210 19:14:33.218274 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6m6c8" event={"ID":"eaa2ffd3-6860-4f1f-9ada-04927d8a81eb","Type":"ContainerStarted","Data":"ca155bc7158ca12a9e9559b7e002e882d3e73f90e376f829ce3ffbfb142818fd"} Dec 10 19:14:36 crc kubenswrapper[4828]: I1210 19:14:36.262574 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-6m6c8" podStartSLOduration=1.335463286 podStartE2EDuration="4.262551961s" podCreationTimestamp="2025-12-10 19:14:32 +0000 UTC" firstStartedPulling="2025-12-10 19:14:33.13505658 +0000 UTC m=+1153.645667595" lastFinishedPulling="2025-12-10 19:14:36.062145265 +0000 UTC m=+1156.572756270" observedRunningTime="2025-12-10 19:14:36.256128103 +0000 UTC m=+1156.766739108" watchObservedRunningTime="2025-12-10 19:14:36.262551961 +0000 UTC m=+1156.773162966" Dec 10 19:14:37 crc kubenswrapper[4828]: I1210 19:14:37.253093 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zlxm5" event={"ID":"3f36c8bd-7eea-416b-bf77-5bca884a3426","Type":"ContainerStarted","Data":"ae9f4eaf12f1a8acec73fcb2df96c41139b7ac25294cbbe5ed6033f9238c0988"} Dec 10 19:14:37 crc kubenswrapper[4828]: I1210 19:14:37.253176 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-zlxm5" podUID="3f36c8bd-7eea-416b-bf77-5bca884a3426" containerName="registry-server" containerID="cri-o://ae9f4eaf12f1a8acec73fcb2df96c41139b7ac25294cbbe5ed6033f9238c0988" gracePeriod=2 Dec 10 19:14:37 crc kubenswrapper[4828]: I1210 19:14:37.255084 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6m6c8" event={"ID":"eaa2ffd3-6860-4f1f-9ada-04927d8a81eb","Type":"ContainerStarted","Data":"777c5e58dc63dea4fa68d33382009627dccba34c202f97f107b0a23e78ac1e2f"} Dec 10 19:14:37 crc kubenswrapper[4828]: I1210 19:14:37.274578 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-zlxm5" podStartSLOduration=2.094511826 podStartE2EDuration="9.27455641s" podCreationTimestamp="2025-12-10 19:14:28 +0000 UTC" firstStartedPulling="2025-12-10 19:14:28.873719893 +0000 UTC m=+1149.384330898" lastFinishedPulling="2025-12-10 19:14:36.053764477 +0000 UTC m=+1156.564375482" observedRunningTime="2025-12-10 19:14:37.265599536 +0000 UTC m=+1157.776210541" watchObservedRunningTime="2025-12-10 19:14:37.27455641 +0000 UTC m=+1157.785167425" Dec 10 19:14:37 crc kubenswrapper[4828]: I1210 19:14:37.699372 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zlxm5" Dec 10 19:14:37 crc kubenswrapper[4828]: I1210 19:14:37.807668 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl24j\" (UniqueName: \"kubernetes.io/projected/3f36c8bd-7eea-416b-bf77-5bca884a3426-kube-api-access-dl24j\") pod \"3f36c8bd-7eea-416b-bf77-5bca884a3426\" (UID: \"3f36c8bd-7eea-416b-bf77-5bca884a3426\") " Dec 10 19:14:37 crc kubenswrapper[4828]: I1210 19:14:37.812763 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f36c8bd-7eea-416b-bf77-5bca884a3426-kube-api-access-dl24j" (OuterVolumeSpecName: "kube-api-access-dl24j") pod "3f36c8bd-7eea-416b-bf77-5bca884a3426" (UID: "3f36c8bd-7eea-416b-bf77-5bca884a3426"). InnerVolumeSpecName "kube-api-access-dl24j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:14:37 crc kubenswrapper[4828]: I1210 19:14:37.909197 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl24j\" (UniqueName: \"kubernetes.io/projected/3f36c8bd-7eea-416b-bf77-5bca884a3426-kube-api-access-dl24j\") on node \"crc\" DevicePath \"\"" Dec 10 19:14:38 crc kubenswrapper[4828]: I1210 19:14:38.262897 4828 generic.go:334] "Generic (PLEG): container finished" podID="3f36c8bd-7eea-416b-bf77-5bca884a3426" containerID="ae9f4eaf12f1a8acec73fcb2df96c41139b7ac25294cbbe5ed6033f9238c0988" exitCode=0 Dec 10 19:14:38 crc kubenswrapper[4828]: I1210 19:14:38.262959 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zlxm5" Dec 10 19:14:38 crc kubenswrapper[4828]: I1210 19:14:38.262965 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zlxm5" event={"ID":"3f36c8bd-7eea-416b-bf77-5bca884a3426","Type":"ContainerDied","Data":"ae9f4eaf12f1a8acec73fcb2df96c41139b7ac25294cbbe5ed6033f9238c0988"} Dec 10 19:14:38 crc kubenswrapper[4828]: I1210 19:14:38.263041 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zlxm5" event={"ID":"3f36c8bd-7eea-416b-bf77-5bca884a3426","Type":"ContainerDied","Data":"8f13fd6253e2da3f22a32fade60ee67b03db9af5203bbd7751d61888a739446f"} Dec 10 19:14:38 crc kubenswrapper[4828]: I1210 19:14:38.263067 4828 scope.go:117] "RemoveContainer" containerID="ae9f4eaf12f1a8acec73fcb2df96c41139b7ac25294cbbe5ed6033f9238c0988" Dec 10 19:14:38 crc kubenswrapper[4828]: I1210 19:14:38.293370 4828 scope.go:117] "RemoveContainer" containerID="ae9f4eaf12f1a8acec73fcb2df96c41139b7ac25294cbbe5ed6033f9238c0988" Dec 10 19:14:38 crc kubenswrapper[4828]: E1210 19:14:38.293896 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae9f4eaf12f1a8acec73fcb2df96c41139b7ac25294cbbe5ed6033f9238c0988\": container with ID starting with ae9f4eaf12f1a8acec73fcb2df96c41139b7ac25294cbbe5ed6033f9238c0988 not found: ID does not exist" containerID="ae9f4eaf12f1a8acec73fcb2df96c41139b7ac25294cbbe5ed6033f9238c0988" Dec 10 19:14:38 crc kubenswrapper[4828]: I1210 19:14:38.293936 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae9f4eaf12f1a8acec73fcb2df96c41139b7ac25294cbbe5ed6033f9238c0988"} err="failed to get container status \"ae9f4eaf12f1a8acec73fcb2df96c41139b7ac25294cbbe5ed6033f9238c0988\": rpc error: code = NotFound desc = could not find container \"ae9f4eaf12f1a8acec73fcb2df96c41139b7ac25294cbbe5ed6033f9238c0988\": container with ID starting with ae9f4eaf12f1a8acec73fcb2df96c41139b7ac25294cbbe5ed6033f9238c0988 not found: ID does not exist" Dec 10 19:14:38 crc kubenswrapper[4828]: I1210 19:14:38.297070 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-zlxm5"] Dec 10 19:14:38 crc kubenswrapper[4828]: I1210 19:14:38.302367 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-zlxm5"] Dec 10 19:14:39 crc kubenswrapper[4828]: I1210 19:14:39.798972 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f36c8bd-7eea-416b-bf77-5bca884a3426" path="/var/lib/kubelet/pods/3f36c8bd-7eea-416b-bf77-5bca884a3426/volumes" Dec 10 19:14:42 crc kubenswrapper[4828]: I1210 19:14:42.511527 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-6m6c8" Dec 10 19:14:42 crc kubenswrapper[4828]: I1210 19:14:42.511907 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-6m6c8" Dec 10 19:14:42 crc kubenswrapper[4828]: I1210 19:14:42.540574 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-6m6c8" Dec 10 19:14:43 crc kubenswrapper[4828]: I1210 19:14:43.333184 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-6m6c8" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.230195 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.230861 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.230916 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.231718 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6e3402faca982d603d83c3dbbde8307326fa3a18260295ae4bfb613477d32c1e"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.231776 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://6e3402faca982d603d83c3dbbde8307326fa3a18260295ae4bfb613477d32c1e" gracePeriod=600 Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.241685 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7"] Dec 10 19:14:51 crc kubenswrapper[4828]: E1210 19:14:51.242035 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f36c8bd-7eea-416b-bf77-5bca884a3426" containerName="registry-server" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.242055 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f36c8bd-7eea-416b-bf77-5bca884a3426" containerName="registry-server" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.242215 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f36c8bd-7eea-416b-bf77-5bca884a3426" containerName="registry-server" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.243303 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.246081 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-c5bzc" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.255791 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7"] Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.321209 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6hjh\" (UniqueName: \"kubernetes.io/projected/53cc41b8-5f56-44aa-9861-60973e96f7b7-kube-api-access-h6hjh\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7\" (UID: \"53cc41b8-5f56-44aa-9861-60973e96f7b7\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.321546 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/53cc41b8-5f56-44aa-9861-60973e96f7b7-util\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7\" (UID: \"53cc41b8-5f56-44aa-9861-60973e96f7b7\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.321632 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/53cc41b8-5f56-44aa-9861-60973e96f7b7-bundle\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7\" (UID: \"53cc41b8-5f56-44aa-9861-60973e96f7b7\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.367703 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="6e3402faca982d603d83c3dbbde8307326fa3a18260295ae4bfb613477d32c1e" exitCode=0 Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.367743 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"6e3402faca982d603d83c3dbbde8307326fa3a18260295ae4bfb613477d32c1e"} Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.367775 4828 scope.go:117] "RemoveContainer" containerID="97c29c3bfdeed7777dc023da2b04e9fc93587d77035f3f054b943162232fe57b" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.423926 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6hjh\" (UniqueName: \"kubernetes.io/projected/53cc41b8-5f56-44aa-9861-60973e96f7b7-kube-api-access-h6hjh\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7\" (UID: \"53cc41b8-5f56-44aa-9861-60973e96f7b7\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.423973 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/53cc41b8-5f56-44aa-9861-60973e96f7b7-util\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7\" (UID: \"53cc41b8-5f56-44aa-9861-60973e96f7b7\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.424009 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/53cc41b8-5f56-44aa-9861-60973e96f7b7-bundle\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7\" (UID: \"53cc41b8-5f56-44aa-9861-60973e96f7b7\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.424479 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/53cc41b8-5f56-44aa-9861-60973e96f7b7-util\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7\" (UID: \"53cc41b8-5f56-44aa-9861-60973e96f7b7\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.424492 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/53cc41b8-5f56-44aa-9861-60973e96f7b7-bundle\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7\" (UID: \"53cc41b8-5f56-44aa-9861-60973e96f7b7\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.442962 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6hjh\" (UniqueName: \"kubernetes.io/projected/53cc41b8-5f56-44aa-9861-60973e96f7b7-kube-api-access-h6hjh\") pod \"0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7\" (UID: \"53cc41b8-5f56-44aa-9861-60973e96f7b7\") " pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.568533 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" Dec 10 19:14:51 crc kubenswrapper[4828]: I1210 19:14:51.963569 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7"] Dec 10 19:14:52 crc kubenswrapper[4828]: I1210 19:14:52.376026 4828 generic.go:334] "Generic (PLEG): container finished" podID="53cc41b8-5f56-44aa-9861-60973e96f7b7" containerID="75e844bfe7f0323cabcd99525e6d821de49583ca9d6fba48765770302d675b8b" exitCode=0 Dec 10 19:14:52 crc kubenswrapper[4828]: I1210 19:14:52.376214 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" event={"ID":"53cc41b8-5f56-44aa-9861-60973e96f7b7","Type":"ContainerDied","Data":"75e844bfe7f0323cabcd99525e6d821de49583ca9d6fba48765770302d675b8b"} Dec 10 19:14:52 crc kubenswrapper[4828]: I1210 19:14:52.376927 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" event={"ID":"53cc41b8-5f56-44aa-9861-60973e96f7b7","Type":"ContainerStarted","Data":"6c5d19f6d9ed91f762082dcd5312aa0982852f4e8c648a6744eafac6c79925fe"} Dec 10 19:14:52 crc kubenswrapper[4828]: I1210 19:14:52.379185 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"04946921ee0aa69a9d56354bf922c69f57770e0a2b673248523db9ef4278fd20"} Dec 10 19:14:53 crc kubenswrapper[4828]: I1210 19:14:53.389978 4828 generic.go:334] "Generic (PLEG): container finished" podID="53cc41b8-5f56-44aa-9861-60973e96f7b7" containerID="9d78e5979f8e92425bcd8728c0a58fb709c63d51eb2bb748f896f58130db8af2" exitCode=0 Dec 10 19:14:53 crc kubenswrapper[4828]: I1210 19:14:53.390065 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" event={"ID":"53cc41b8-5f56-44aa-9861-60973e96f7b7","Type":"ContainerDied","Data":"9d78e5979f8e92425bcd8728c0a58fb709c63d51eb2bb748f896f58130db8af2"} Dec 10 19:14:54 crc kubenswrapper[4828]: I1210 19:14:54.401129 4828 generic.go:334] "Generic (PLEG): container finished" podID="53cc41b8-5f56-44aa-9861-60973e96f7b7" containerID="adb47ffa48c10e49204113251303b3a51d839b3ea031fdf96bb34aaf8e2ad89d" exitCode=0 Dec 10 19:14:54 crc kubenswrapper[4828]: I1210 19:14:54.401233 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" event={"ID":"53cc41b8-5f56-44aa-9861-60973e96f7b7","Type":"ContainerDied","Data":"adb47ffa48c10e49204113251303b3a51d839b3ea031fdf96bb34aaf8e2ad89d"} Dec 10 19:14:55 crc kubenswrapper[4828]: I1210 19:14:55.719747 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" Dec 10 19:14:55 crc kubenswrapper[4828]: I1210 19:14:55.795381 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/53cc41b8-5f56-44aa-9861-60973e96f7b7-util\") pod \"53cc41b8-5f56-44aa-9861-60973e96f7b7\" (UID: \"53cc41b8-5f56-44aa-9861-60973e96f7b7\") " Dec 10 19:14:55 crc kubenswrapper[4828]: I1210 19:14:55.795433 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/53cc41b8-5f56-44aa-9861-60973e96f7b7-bundle\") pod \"53cc41b8-5f56-44aa-9861-60973e96f7b7\" (UID: \"53cc41b8-5f56-44aa-9861-60973e96f7b7\") " Dec 10 19:14:55 crc kubenswrapper[4828]: I1210 19:14:55.796317 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6hjh\" (UniqueName: \"kubernetes.io/projected/53cc41b8-5f56-44aa-9861-60973e96f7b7-kube-api-access-h6hjh\") pod \"53cc41b8-5f56-44aa-9861-60973e96f7b7\" (UID: \"53cc41b8-5f56-44aa-9861-60973e96f7b7\") " Dec 10 19:14:55 crc kubenswrapper[4828]: I1210 19:14:55.796967 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53cc41b8-5f56-44aa-9861-60973e96f7b7-bundle" (OuterVolumeSpecName: "bundle") pod "53cc41b8-5f56-44aa-9861-60973e96f7b7" (UID: "53cc41b8-5f56-44aa-9861-60973e96f7b7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:14:55 crc kubenswrapper[4828]: I1210 19:14:55.802529 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53cc41b8-5f56-44aa-9861-60973e96f7b7-kube-api-access-h6hjh" (OuterVolumeSpecName: "kube-api-access-h6hjh") pod "53cc41b8-5f56-44aa-9861-60973e96f7b7" (UID: "53cc41b8-5f56-44aa-9861-60973e96f7b7"). InnerVolumeSpecName "kube-api-access-h6hjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:14:55 crc kubenswrapper[4828]: I1210 19:14:55.808446 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53cc41b8-5f56-44aa-9861-60973e96f7b7-util" (OuterVolumeSpecName: "util") pod "53cc41b8-5f56-44aa-9861-60973e96f7b7" (UID: "53cc41b8-5f56-44aa-9861-60973e96f7b7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:14:55 crc kubenswrapper[4828]: I1210 19:14:55.898367 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6hjh\" (UniqueName: \"kubernetes.io/projected/53cc41b8-5f56-44aa-9861-60973e96f7b7-kube-api-access-h6hjh\") on node \"crc\" DevicePath \"\"" Dec 10 19:14:55 crc kubenswrapper[4828]: I1210 19:14:55.898411 4828 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/53cc41b8-5f56-44aa-9861-60973e96f7b7-util\") on node \"crc\" DevicePath \"\"" Dec 10 19:14:55 crc kubenswrapper[4828]: I1210 19:14:55.898426 4828 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/53cc41b8-5f56-44aa-9861-60973e96f7b7-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:14:56 crc kubenswrapper[4828]: I1210 19:14:56.420313 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" event={"ID":"53cc41b8-5f56-44aa-9861-60973e96f7b7","Type":"ContainerDied","Data":"6c5d19f6d9ed91f762082dcd5312aa0982852f4e8c648a6744eafac6c79925fe"} Dec 10 19:14:56 crc kubenswrapper[4828]: I1210 19:14:56.420662 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c5d19f6d9ed91f762082dcd5312aa0982852f4e8c648a6744eafac6c79925fe" Dec 10 19:14:56 crc kubenswrapper[4828]: I1210 19:14:56.420390 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.144616 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k"] Dec 10 19:15:00 crc kubenswrapper[4828]: E1210 19:15:00.145887 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53cc41b8-5f56-44aa-9861-60973e96f7b7" containerName="extract" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.145922 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="53cc41b8-5f56-44aa-9861-60973e96f7b7" containerName="extract" Dec 10 19:15:00 crc kubenswrapper[4828]: E1210 19:15:00.145974 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53cc41b8-5f56-44aa-9861-60973e96f7b7" containerName="util" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.145991 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="53cc41b8-5f56-44aa-9861-60973e96f7b7" containerName="util" Dec 10 19:15:00 crc kubenswrapper[4828]: E1210 19:15:00.146027 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53cc41b8-5f56-44aa-9861-60973e96f7b7" containerName="pull" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.146047 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="53cc41b8-5f56-44aa-9861-60973e96f7b7" containerName="pull" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.146451 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="53cc41b8-5f56-44aa-9861-60973e96f7b7" containerName="extract" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.147766 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.150379 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.150605 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.154641 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k"] Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.170893 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtldf\" (UniqueName: \"kubernetes.io/projected/13848851-6a8f-48ce-a64e-870a6d47993d-kube-api-access-vtldf\") pod \"collect-profiles-29423235-pg58k\" (UID: \"13848851-6a8f-48ce-a64e-870a6d47993d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.171353 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13848851-6a8f-48ce-a64e-870a6d47993d-secret-volume\") pod \"collect-profiles-29423235-pg58k\" (UID: \"13848851-6a8f-48ce-a64e-870a6d47993d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.171414 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13848851-6a8f-48ce-a64e-870a6d47993d-config-volume\") pod \"collect-profiles-29423235-pg58k\" (UID: \"13848851-6a8f-48ce-a64e-870a6d47993d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.273507 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13848851-6a8f-48ce-a64e-870a6d47993d-config-volume\") pod \"collect-profiles-29423235-pg58k\" (UID: \"13848851-6a8f-48ce-a64e-870a6d47993d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.273616 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtldf\" (UniqueName: \"kubernetes.io/projected/13848851-6a8f-48ce-a64e-870a6d47993d-kube-api-access-vtldf\") pod \"collect-profiles-29423235-pg58k\" (UID: \"13848851-6a8f-48ce-a64e-870a6d47993d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.273767 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13848851-6a8f-48ce-a64e-870a6d47993d-secret-volume\") pod \"collect-profiles-29423235-pg58k\" (UID: \"13848851-6a8f-48ce-a64e-870a6d47993d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.274891 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13848851-6a8f-48ce-a64e-870a6d47993d-config-volume\") pod \"collect-profiles-29423235-pg58k\" (UID: \"13848851-6a8f-48ce-a64e-870a6d47993d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.282367 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13848851-6a8f-48ce-a64e-870a6d47993d-secret-volume\") pod \"collect-profiles-29423235-pg58k\" (UID: \"13848851-6a8f-48ce-a64e-870a6d47993d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.290404 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtldf\" (UniqueName: \"kubernetes.io/projected/13848851-6a8f-48ce-a64e-870a6d47993d-kube-api-access-vtldf\") pod \"collect-profiles-29423235-pg58k\" (UID: \"13848851-6a8f-48ce-a64e-870a6d47993d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.469357 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k" Dec 10 19:15:00 crc kubenswrapper[4828]: I1210 19:15:00.854149 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k"] Dec 10 19:15:01 crc kubenswrapper[4828]: I1210 19:15:01.459456 4828 generic.go:334] "Generic (PLEG): container finished" podID="13848851-6a8f-48ce-a64e-870a6d47993d" containerID="f227a7ac24dbd3b95a332f31c411345447e280b0d5ac94889b590869ae2e8c0b" exitCode=0 Dec 10 19:15:01 crc kubenswrapper[4828]: I1210 19:15:01.459550 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k" event={"ID":"13848851-6a8f-48ce-a64e-870a6d47993d","Type":"ContainerDied","Data":"f227a7ac24dbd3b95a332f31c411345447e280b0d5ac94889b590869ae2e8c0b"} Dec 10 19:15:01 crc kubenswrapper[4828]: I1210 19:15:01.459763 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k" event={"ID":"13848851-6a8f-48ce-a64e-870a6d47993d","Type":"ContainerStarted","Data":"03975e8888eb614a7f15db3aa7b40e623723a823eac1247cf887adce7b7f2742"} Dec 10 19:15:02 crc kubenswrapper[4828]: I1210 19:15:02.822908 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k" Dec 10 19:15:02 crc kubenswrapper[4828]: I1210 19:15:02.935264 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtldf\" (UniqueName: \"kubernetes.io/projected/13848851-6a8f-48ce-a64e-870a6d47993d-kube-api-access-vtldf\") pod \"13848851-6a8f-48ce-a64e-870a6d47993d\" (UID: \"13848851-6a8f-48ce-a64e-870a6d47993d\") " Dec 10 19:15:02 crc kubenswrapper[4828]: I1210 19:15:02.935325 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13848851-6a8f-48ce-a64e-870a6d47993d-config-volume\") pod \"13848851-6a8f-48ce-a64e-870a6d47993d\" (UID: \"13848851-6a8f-48ce-a64e-870a6d47993d\") " Dec 10 19:15:02 crc kubenswrapper[4828]: I1210 19:15:02.935474 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13848851-6a8f-48ce-a64e-870a6d47993d-secret-volume\") pod \"13848851-6a8f-48ce-a64e-870a6d47993d\" (UID: \"13848851-6a8f-48ce-a64e-870a6d47993d\") " Dec 10 19:15:02 crc kubenswrapper[4828]: I1210 19:15:02.936254 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13848851-6a8f-48ce-a64e-870a6d47993d-config-volume" (OuterVolumeSpecName: "config-volume") pod "13848851-6a8f-48ce-a64e-870a6d47993d" (UID: "13848851-6a8f-48ce-a64e-870a6d47993d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:15:02 crc kubenswrapper[4828]: I1210 19:15:02.941719 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13848851-6a8f-48ce-a64e-870a6d47993d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "13848851-6a8f-48ce-a64e-870a6d47993d" (UID: "13848851-6a8f-48ce-a64e-870a6d47993d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:15:02 crc kubenswrapper[4828]: I1210 19:15:02.941881 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13848851-6a8f-48ce-a64e-870a6d47993d-kube-api-access-vtldf" (OuterVolumeSpecName: "kube-api-access-vtldf") pod "13848851-6a8f-48ce-a64e-870a6d47993d" (UID: "13848851-6a8f-48ce-a64e-870a6d47993d"). InnerVolumeSpecName "kube-api-access-vtldf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:15:03 crc kubenswrapper[4828]: I1210 19:15:03.037515 4828 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13848851-6a8f-48ce-a64e-870a6d47993d-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:03 crc kubenswrapper[4828]: I1210 19:15:03.037551 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtldf\" (UniqueName: \"kubernetes.io/projected/13848851-6a8f-48ce-a64e-870a6d47993d-kube-api-access-vtldf\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:03 crc kubenswrapper[4828]: I1210 19:15:03.037567 4828 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13848851-6a8f-48ce-a64e-870a6d47993d-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 19:15:03 crc kubenswrapper[4828]: I1210 19:15:03.226739 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6b7b77566b-7lqnm"] Dec 10 19:15:03 crc kubenswrapper[4828]: E1210 19:15:03.227162 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13848851-6a8f-48ce-a64e-870a6d47993d" containerName="collect-profiles" Dec 10 19:15:03 crc kubenswrapper[4828]: I1210 19:15:03.227181 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="13848851-6a8f-48ce-a64e-870a6d47993d" containerName="collect-profiles" Dec 10 19:15:03 crc kubenswrapper[4828]: I1210 19:15:03.227491 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="13848851-6a8f-48ce-a64e-870a6d47993d" containerName="collect-profiles" Dec 10 19:15:03 crc kubenswrapper[4828]: I1210 19:15:03.228210 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-7lqnm" Dec 10 19:15:03 crc kubenswrapper[4828]: I1210 19:15:03.234985 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-4xrld" Dec 10 19:15:03 crc kubenswrapper[4828]: I1210 19:15:03.252207 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6b7b77566b-7lqnm"] Dec 10 19:15:03 crc kubenswrapper[4828]: I1210 19:15:03.341268 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv6fj\" (UniqueName: \"kubernetes.io/projected/273ed82b-72f4-4fa2-b9af-e3269d82e69e-kube-api-access-kv6fj\") pod \"openstack-operator-controller-operator-6b7b77566b-7lqnm\" (UID: \"273ed82b-72f4-4fa2-b9af-e3269d82e69e\") " pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-7lqnm" Dec 10 19:15:03 crc kubenswrapper[4828]: I1210 19:15:03.442820 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv6fj\" (UniqueName: \"kubernetes.io/projected/273ed82b-72f4-4fa2-b9af-e3269d82e69e-kube-api-access-kv6fj\") pod \"openstack-operator-controller-operator-6b7b77566b-7lqnm\" (UID: \"273ed82b-72f4-4fa2-b9af-e3269d82e69e\") " pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-7lqnm" Dec 10 19:15:03 crc kubenswrapper[4828]: I1210 19:15:03.458687 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv6fj\" (UniqueName: \"kubernetes.io/projected/273ed82b-72f4-4fa2-b9af-e3269d82e69e-kube-api-access-kv6fj\") pod \"openstack-operator-controller-operator-6b7b77566b-7lqnm\" (UID: \"273ed82b-72f4-4fa2-b9af-e3269d82e69e\") " pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-7lqnm" Dec 10 19:15:03 crc kubenswrapper[4828]: I1210 19:15:03.476060 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k" event={"ID":"13848851-6a8f-48ce-a64e-870a6d47993d","Type":"ContainerDied","Data":"03975e8888eb614a7f15db3aa7b40e623723a823eac1247cf887adce7b7f2742"} Dec 10 19:15:03 crc kubenswrapper[4828]: I1210 19:15:03.476104 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03975e8888eb614a7f15db3aa7b40e623723a823eac1247cf887adce7b7f2742" Dec 10 19:15:03 crc kubenswrapper[4828]: I1210 19:15:03.476163 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k" Dec 10 19:15:03 crc kubenswrapper[4828]: I1210 19:15:03.555520 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-7lqnm" Dec 10 19:15:03 crc kubenswrapper[4828]: I1210 19:15:03.992250 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6b7b77566b-7lqnm"] Dec 10 19:15:03 crc kubenswrapper[4828]: W1210 19:15:03.995306 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod273ed82b_72f4_4fa2_b9af_e3269d82e69e.slice/crio-e0748867cca45bcf34ef743b4b6a9a57c295843bad65a5db1a14cd007d5cbf73 WatchSource:0}: Error finding container e0748867cca45bcf34ef743b4b6a9a57c295843bad65a5db1a14cd007d5cbf73: Status 404 returned error can't find the container with id e0748867cca45bcf34ef743b4b6a9a57c295843bad65a5db1a14cd007d5cbf73 Dec 10 19:15:04 crc kubenswrapper[4828]: I1210 19:15:04.486240 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-7lqnm" event={"ID":"273ed82b-72f4-4fa2-b9af-e3269d82e69e","Type":"ContainerStarted","Data":"e0748867cca45bcf34ef743b4b6a9a57c295843bad65a5db1a14cd007d5cbf73"} Dec 10 19:15:12 crc kubenswrapper[4828]: I1210 19:15:12.545843 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-7lqnm" event={"ID":"273ed82b-72f4-4fa2-b9af-e3269d82e69e","Type":"ContainerStarted","Data":"614713eaae62ae399eb42bdfb24314c70be62fde32e03b09e61436c17b77a107"} Dec 10 19:15:12 crc kubenswrapper[4828]: I1210 19:15:12.546501 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-7lqnm" Dec 10 19:15:12 crc kubenswrapper[4828]: I1210 19:15:12.577095 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-7lqnm" podStartSLOduration=1.6338236419999999 podStartE2EDuration="9.577077322s" podCreationTimestamp="2025-12-10 19:15:03 +0000 UTC" firstStartedPulling="2025-12-10 19:15:03.998029461 +0000 UTC m=+1184.508640466" lastFinishedPulling="2025-12-10 19:15:11.941283151 +0000 UTC m=+1192.451894146" observedRunningTime="2025-12-10 19:15:12.570467279 +0000 UTC m=+1193.081078294" watchObservedRunningTime="2025-12-10 19:15:12.577077322 +0000 UTC m=+1193.087688327" Dec 10 19:15:23 crc kubenswrapper[4828]: I1210 19:15:23.560699 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6b7b77566b-7lqnm" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.638092 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-pcfhr"] Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.640310 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-pcfhr" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.643026 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-7j4ls" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.644615 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-pcfhr"] Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.652637 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-cfhjs"] Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.655011 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-cfhjs" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.657969 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-lz6k5" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.688196 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-wtcz4"] Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.689578 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-wtcz4" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.695481 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-cfhjs"] Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.698985 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-nn5v5" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.703115 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-vhf9j"] Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.704393 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vhf9j" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.707759 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-gjjqz" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.709952 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-wtcz4"] Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.718036 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-vhf9j"] Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.724636 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mmqkg"] Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.726111 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mmqkg" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.727404 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7c98\" (UniqueName: \"kubernetes.io/projected/b0544042-b6b7-4a9d-884a-12859ddea597-kube-api-access-d7c98\") pod \"barbican-operator-controller-manager-7d9dfd778-pcfhr\" (UID: \"b0544042-b6b7-4a9d-884a-12859ddea597\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-pcfhr" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.745557 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-m82kn" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.906324 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m9pq\" (UniqueName: \"kubernetes.io/projected/1f3d156d-ddc6-47a2-9c40-5b675e535b37-kube-api-access-8m9pq\") pod \"glance-operator-controller-manager-5697bb5779-vhf9j\" (UID: \"1f3d156d-ddc6-47a2-9c40-5b675e535b37\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vhf9j" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.906430 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w428\" (UniqueName: \"kubernetes.io/projected/bd786401-638a-4830-8a28-cc01b03d311d-kube-api-access-6w428\") pod \"designate-operator-controller-manager-697fb699cf-wtcz4\" (UID: \"bd786401-638a-4830-8a28-cc01b03d311d\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-wtcz4" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.906464 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5rd2\" (UniqueName: \"kubernetes.io/projected/bad66884-aa3d-4683-abc6-2387140d407f-kube-api-access-s5rd2\") pod \"heat-operator-controller-manager-5f64f6f8bb-mmqkg\" (UID: \"bad66884-aa3d-4683-abc6-2387140d407f\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mmqkg" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.906496 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7c98\" (UniqueName: \"kubernetes.io/projected/b0544042-b6b7-4a9d-884a-12859ddea597-kube-api-access-d7c98\") pod \"barbican-operator-controller-manager-7d9dfd778-pcfhr\" (UID: \"b0544042-b6b7-4a9d-884a-12859ddea597\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-pcfhr" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.906515 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrxgf\" (UniqueName: \"kubernetes.io/projected/8991af61-9866-4ae9-b7a7-7545833c4485-kube-api-access-lrxgf\") pod \"cinder-operator-controller-manager-6c677c69b-cfhjs\" (UID: \"8991af61-9866-4ae9-b7a7-7545833c4485\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-cfhjs" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.921086 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mmqkg"] Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.928738 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt"] Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.930713 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.933215 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-nqqwl" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.935732 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg"] Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.936379 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.937540 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.940233 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-7dvw2" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.946849 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7c98\" (UniqueName: \"kubernetes.io/projected/b0544042-b6b7-4a9d-884a-12859ddea597-kube-api-access-d7c98\") pod \"barbican-operator-controller-manager-7d9dfd778-pcfhr\" (UID: \"b0544042-b6b7-4a9d-884a-12859ddea597\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-pcfhr" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.956100 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-f7cn8"] Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.957752 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ftlx"] Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.959061 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ftlx" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.959360 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-f7cn8" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.964352 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-pcfhr" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.966467 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-d98v4" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.979668 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-xwsnx" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.991713 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-njhhp"] Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.993716 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-njhhp" Dec 10 19:15:49 crc kubenswrapper[4828]: I1210 19:15:49.997180 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-gqrjw" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.007660 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert\") pod \"infra-operator-controller-manager-78d48bff9d-x27kt\" (UID: \"55bb5bf6-cc44-43f9-84e5-1137401fd188\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.007749 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w428\" (UniqueName: \"kubernetes.io/projected/bd786401-638a-4830-8a28-cc01b03d311d-kube-api-access-6w428\") pod \"designate-operator-controller-manager-697fb699cf-wtcz4\" (UID: \"bd786401-638a-4830-8a28-cc01b03d311d\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-wtcz4" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.007786 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5rd2\" (UniqueName: \"kubernetes.io/projected/bad66884-aa3d-4683-abc6-2387140d407f-kube-api-access-s5rd2\") pod \"heat-operator-controller-manager-5f64f6f8bb-mmqkg\" (UID: \"bad66884-aa3d-4683-abc6-2387140d407f\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mmqkg" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.007829 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrxgf\" (UniqueName: \"kubernetes.io/projected/8991af61-9866-4ae9-b7a7-7545833c4485-kube-api-access-lrxgf\") pod \"cinder-operator-controller-manager-6c677c69b-cfhjs\" (UID: \"8991af61-9866-4ae9-b7a7-7545833c4485\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-cfhjs" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.007864 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m9pq\" (UniqueName: \"kubernetes.io/projected/1f3d156d-ddc6-47a2-9c40-5b675e535b37-kube-api-access-8m9pq\") pod \"glance-operator-controller-manager-5697bb5779-vhf9j\" (UID: \"1f3d156d-ddc6-47a2-9c40-5b675e535b37\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vhf9j" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.007882 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blcrk\" (UniqueName: \"kubernetes.io/projected/55bb5bf6-cc44-43f9-84e5-1137401fd188-kube-api-access-blcrk\") pod \"infra-operator-controller-manager-78d48bff9d-x27kt\" (UID: \"55bb5bf6-cc44-43f9-84e5-1137401fd188\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.013081 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6bqjc"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.014389 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6bqjc" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.024022 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-4jg6j" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.035050 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-smtjq"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.036778 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-smtjq" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.041617 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-9dfrp" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.042228 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5rd2\" (UniqueName: \"kubernetes.io/projected/bad66884-aa3d-4683-abc6-2387140d407f-kube-api-access-s5rd2\") pod \"heat-operator-controller-manager-5f64f6f8bb-mmqkg\" (UID: \"bad66884-aa3d-4683-abc6-2387140d407f\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mmqkg" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.043118 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrxgf\" (UniqueName: \"kubernetes.io/projected/8991af61-9866-4ae9-b7a7-7545833c4485-kube-api-access-lrxgf\") pod \"cinder-operator-controller-manager-6c677c69b-cfhjs\" (UID: \"8991af61-9866-4ae9-b7a7-7545833c4485\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-cfhjs" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.050933 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m9pq\" (UniqueName: \"kubernetes.io/projected/1f3d156d-ddc6-47a2-9c40-5b675e535b37-kube-api-access-8m9pq\") pod \"glance-operator-controller-manager-5697bb5779-vhf9j\" (UID: \"1f3d156d-ddc6-47a2-9c40-5b675e535b37\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vhf9j" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.054725 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w428\" (UniqueName: \"kubernetes.io/projected/bd786401-638a-4830-8a28-cc01b03d311d-kube-api-access-6w428\") pod \"designate-operator-controller-manager-697fb699cf-wtcz4\" (UID: \"bd786401-638a-4830-8a28-cc01b03d311d\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-wtcz4" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.057767 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vhf9j" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.066636 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-j64bx"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.067951 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-j64bx" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.072210 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-xld94" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.087839 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-79p7l"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.090108 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-79p7l" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.090972 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mmqkg" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.091635 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-g2sfc" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.106117 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-dxcl8"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.107928 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-dxcl8" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.112700 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwmjr\" (UniqueName: \"kubernetes.io/projected/e87fb7c5-676f-4549-82e3-2696306f651b-kube-api-access-bwmjr\") pod \"horizon-operator-controller-manager-68c6d99b8f-s5jjg\" (UID: \"e87fb7c5-676f-4549-82e3-2696306f651b\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.112820 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q6sn\" (UniqueName: \"kubernetes.io/projected/23431602-592f-4c26-a5ab-1ada81fda48a-kube-api-access-7q6sn\") pod \"mariadb-operator-controller-manager-79c8c4686c-6bqjc\" (UID: \"23431602-592f-4c26-a5ab-1ada81fda48a\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6bqjc" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.112852 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfd26\" (UniqueName: \"kubernetes.io/projected/be987ffc-f75a-4364-ac29-37fbc9bb7200-kube-api-access-nfd26\") pod \"keystone-operator-controller-manager-7765d96ddf-4ftlx\" (UID: \"be987ffc-f75a-4364-ac29-37fbc9bb7200\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ftlx" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.112884 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blcrk\" (UniqueName: \"kubernetes.io/projected/55bb5bf6-cc44-43f9-84e5-1137401fd188-kube-api-access-blcrk\") pod \"infra-operator-controller-manager-78d48bff9d-x27kt\" (UID: \"55bb5bf6-cc44-43f9-84e5-1137401fd188\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.112990 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw6cg\" (UniqueName: \"kubernetes.io/projected/85a6c448-c39f-4ec5-b1f7-4f7ff1bc43a0-kube-api-access-zw6cg\") pod \"ironic-operator-controller-manager-967d97867-f7cn8\" (UID: \"85a6c448-c39f-4ec5-b1f7-4f7ff1bc43a0\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-f7cn8" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.113055 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert\") pod \"infra-operator-controller-manager-78d48bff9d-x27kt\" (UID: \"55bb5bf6-cc44-43f9-84e5-1137401fd188\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.113218 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjvj5\" (UniqueName: \"kubernetes.io/projected/ab0dbc03-b62b-40c9-98b9-7bb192cddff3-kube-api-access-sjvj5\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-smtjq\" (UID: \"ab0dbc03-b62b-40c9-98b9-7bb192cddff3\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-smtjq" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.113288 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwp2h\" (UniqueName: \"kubernetes.io/projected/8f635d23-992f-4a05-9896-c38d41ec853d-kube-api-access-kwp2h\") pod \"manila-operator-controller-manager-5b5fd79c9c-njhhp\" (UID: \"8f635d23-992f-4a05-9896-c38d41ec853d\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-njhhp" Dec 10 19:15:50 crc kubenswrapper[4828]: E1210 19:15:50.113758 4828 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 19:15:50 crc kubenswrapper[4828]: E1210 19:15:50.113823 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert podName:55bb5bf6-cc44-43f9-84e5-1137401fd188 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:50.613787378 +0000 UTC m=+1231.124398383 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert") pod "infra-operator-controller-manager-78d48bff9d-x27kt" (UID: "55bb5bf6-cc44-43f9-84e5-1137401fd188") : secret "infra-operator-webhook-server-cert" not found Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.116359 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-dd85v" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.132573 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.135577 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.139186 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blcrk\" (UniqueName: \"kubernetes.io/projected/55bb5bf6-cc44-43f9-84e5-1137401fd188-kube-api-access-blcrk\") pod \"infra-operator-controller-manager-78d48bff9d-x27kt\" (UID: \"55bb5bf6-cc44-43f9-84e5-1137401fd188\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.141046 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-gf7r5" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.143552 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.163294 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.173079 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-f7cn8"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.191059 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-j64bx"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.210654 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.214836 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q6sn\" (UniqueName: \"kubernetes.io/projected/23431602-592f-4c26-a5ab-1ada81fda48a-kube-api-access-7q6sn\") pod \"mariadb-operator-controller-manager-79c8c4686c-6bqjc\" (UID: \"23431602-592f-4c26-a5ab-1ada81fda48a\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6bqjc" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.214864 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfd26\" (UniqueName: \"kubernetes.io/projected/be987ffc-f75a-4364-ac29-37fbc9bb7200-kube-api-access-nfd26\") pod \"keystone-operator-controller-manager-7765d96ddf-4ftlx\" (UID: \"be987ffc-f75a-4364-ac29-37fbc9bb7200\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ftlx" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.214906 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw6cg\" (UniqueName: \"kubernetes.io/projected/85a6c448-c39f-4ec5-b1f7-4f7ff1bc43a0-kube-api-access-zw6cg\") pod \"ironic-operator-controller-manager-967d97867-f7cn8\" (UID: \"85a6c448-c39f-4ec5-b1f7-4f7ff1bc43a0\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-f7cn8" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.214935 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqw2j\" (UniqueName: \"kubernetes.io/projected/70814aec-fe59-4dc6-a085-c5a5da7b93a5-kube-api-access-tqw2j\") pod \"nova-operator-controller-manager-697bc559fc-j64bx\" (UID: \"70814aec-fe59-4dc6-a085-c5a5da7b93a5\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-j64bx" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.214972 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvqcg\" (UniqueName: \"kubernetes.io/projected/4614b8b1-2082-4b05-8473-404f0f82772b-kube-api-access-qvqcg\") pod \"openstack-baremetal-operator-controller-manager-84b575879fk7hbh\" (UID: \"4614b8b1-2082-4b05-8473-404f0f82772b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.215000 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt46s\" (UniqueName: \"kubernetes.io/projected/4246d7f6-cfb7-4d1d-aac0-6ba9db04a50b-kube-api-access-nt46s\") pod \"ovn-operator-controller-manager-b6456fdb6-dxcl8\" (UID: \"4246d7f6-cfb7-4d1d-aac0-6ba9db04a50b\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-dxcl8" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.215030 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fk7hbh\" (UID: \"4614b8b1-2082-4b05-8473-404f0f82772b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.215058 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppxxl\" (UniqueName: \"kubernetes.io/projected/be915c9b-8dfe-4aab-8e74-5375e365e623-kube-api-access-ppxxl\") pod \"octavia-operator-controller-manager-998648c74-79p7l\" (UID: \"be915c9b-8dfe-4aab-8e74-5375e365e623\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-79p7l" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.215086 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjvj5\" (UniqueName: \"kubernetes.io/projected/ab0dbc03-b62b-40c9-98b9-7bb192cddff3-kube-api-access-sjvj5\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-smtjq\" (UID: \"ab0dbc03-b62b-40c9-98b9-7bb192cddff3\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-smtjq" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.215121 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwp2h\" (UniqueName: \"kubernetes.io/projected/8f635d23-992f-4a05-9896-c38d41ec853d-kube-api-access-kwp2h\") pod \"manila-operator-controller-manager-5b5fd79c9c-njhhp\" (UID: \"8f635d23-992f-4a05-9896-c38d41ec853d\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-njhhp" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.215153 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwmjr\" (UniqueName: \"kubernetes.io/projected/e87fb7c5-676f-4549-82e3-2696306f651b-kube-api-access-bwmjr\") pod \"horizon-operator-controller-manager-68c6d99b8f-s5jjg\" (UID: \"e87fb7c5-676f-4549-82e3-2696306f651b\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.239109 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-smtjq"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.245588 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfd26\" (UniqueName: \"kubernetes.io/projected/be987ffc-f75a-4364-ac29-37fbc9bb7200-kube-api-access-nfd26\") pod \"keystone-operator-controller-manager-7765d96ddf-4ftlx\" (UID: \"be987ffc-f75a-4364-ac29-37fbc9bb7200\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ftlx" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.252050 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q6sn\" (UniqueName: \"kubernetes.io/projected/23431602-592f-4c26-a5ab-1ada81fda48a-kube-api-access-7q6sn\") pod \"mariadb-operator-controller-manager-79c8c4686c-6bqjc\" (UID: \"23431602-592f-4c26-a5ab-1ada81fda48a\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6bqjc" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.257526 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwmjr\" (UniqueName: \"kubernetes.io/projected/e87fb7c5-676f-4549-82e3-2696306f651b-kube-api-access-bwmjr\") pod \"horizon-operator-controller-manager-68c6d99b8f-s5jjg\" (UID: \"e87fb7c5-676f-4549-82e3-2696306f651b\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.257783 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwp2h\" (UniqueName: \"kubernetes.io/projected/8f635d23-992f-4a05-9896-c38d41ec853d-kube-api-access-kwp2h\") pod \"manila-operator-controller-manager-5b5fd79c9c-njhhp\" (UID: \"8f635d23-992f-4a05-9896-c38d41ec853d\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-njhhp" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.257789 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw6cg\" (UniqueName: \"kubernetes.io/projected/85a6c448-c39f-4ec5-b1f7-4f7ff1bc43a0-kube-api-access-zw6cg\") pod \"ironic-operator-controller-manager-967d97867-f7cn8\" (UID: \"85a6c448-c39f-4ec5-b1f7-4f7ff1bc43a0\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-f7cn8" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.258078 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjvj5\" (UniqueName: \"kubernetes.io/projected/ab0dbc03-b62b-40c9-98b9-7bb192cddff3-kube-api-access-sjvj5\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-smtjq\" (UID: \"ab0dbc03-b62b-40c9-98b9-7bb192cddff3\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-smtjq" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.268527 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-f7cn8" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.270209 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-njhhp"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.280389 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-cfhjs" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.316310 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqw2j\" (UniqueName: \"kubernetes.io/projected/70814aec-fe59-4dc6-a085-c5a5da7b93a5-kube-api-access-tqw2j\") pod \"nova-operator-controller-manager-697bc559fc-j64bx\" (UID: \"70814aec-fe59-4dc6-a085-c5a5da7b93a5\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-j64bx" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.316357 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvqcg\" (UniqueName: \"kubernetes.io/projected/4614b8b1-2082-4b05-8473-404f0f82772b-kube-api-access-qvqcg\") pod \"openstack-baremetal-operator-controller-manager-84b575879fk7hbh\" (UID: \"4614b8b1-2082-4b05-8473-404f0f82772b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.316381 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt46s\" (UniqueName: \"kubernetes.io/projected/4246d7f6-cfb7-4d1d-aac0-6ba9db04a50b-kube-api-access-nt46s\") pod \"ovn-operator-controller-manager-b6456fdb6-dxcl8\" (UID: \"4246d7f6-cfb7-4d1d-aac0-6ba9db04a50b\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-dxcl8" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.316413 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fk7hbh\" (UID: \"4614b8b1-2082-4b05-8473-404f0f82772b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.316446 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppxxl\" (UniqueName: \"kubernetes.io/projected/be915c9b-8dfe-4aab-8e74-5375e365e623-kube-api-access-ppxxl\") pod \"octavia-operator-controller-manager-998648c74-79p7l\" (UID: \"be915c9b-8dfe-4aab-8e74-5375e365e623\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-79p7l" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.317254 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-dxcl8"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.327629 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-wtcz4" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.329973 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-88stb"] Dec 10 19:15:50 crc kubenswrapper[4828]: E1210 19:15:50.330104 4828 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:15:50 crc kubenswrapper[4828]: E1210 19:15:50.330150 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert podName:4614b8b1-2082-4b05-8473-404f0f82772b nodeName:}" failed. No retries permitted until 2025-12-10 19:15:50.830135737 +0000 UTC m=+1231.340746742 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fk7hbh" (UID: "4614b8b1-2082-4b05-8473-404f0f82772b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.332712 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-88stb" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.333162 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6bqjc" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.333696 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-njhhp" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.338267 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-ltgsq" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.348638 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-smtjq" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.352325 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-88stb"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.355442 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppxxl\" (UniqueName: \"kubernetes.io/projected/be915c9b-8dfe-4aab-8e74-5375e365e623-kube-api-access-ppxxl\") pod \"octavia-operator-controller-manager-998648c74-79p7l\" (UID: \"be915c9b-8dfe-4aab-8e74-5375e365e623\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-79p7l" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.377063 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt46s\" (UniqueName: \"kubernetes.io/projected/4246d7f6-cfb7-4d1d-aac0-6ba9db04a50b-kube-api-access-nt46s\") pod \"ovn-operator-controller-manager-b6456fdb6-dxcl8\" (UID: \"4246d7f6-cfb7-4d1d-aac0-6ba9db04a50b\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-dxcl8" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.383078 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ftlx"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.384242 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-79p7l" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.394750 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvqcg\" (UniqueName: \"kubernetes.io/projected/4614b8b1-2082-4b05-8473-404f0f82772b-kube-api-access-qvqcg\") pod \"openstack-baremetal-operator-controller-manager-84b575879fk7hbh\" (UID: \"4614b8b1-2082-4b05-8473-404f0f82772b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.395246 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqw2j\" (UniqueName: \"kubernetes.io/projected/70814aec-fe59-4dc6-a085-c5a5da7b93a5-kube-api-access-tqw2j\") pod \"nova-operator-controller-manager-697bc559fc-j64bx\" (UID: \"70814aec-fe59-4dc6-a085-c5a5da7b93a5\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-j64bx" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.401252 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.411765 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-dxcl8" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.416983 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6bqjc"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.418970 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lpxm\" (UniqueName: \"kubernetes.io/projected/3c945977-c07e-4ef4-9370-d2fa1264eee0-kube-api-access-2lpxm\") pod \"placement-operator-controller-manager-78f8948974-88stb\" (UID: \"3c945977-c07e-4ef4-9370-d2fa1264eee0\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-88stb" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.438910 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-pf44z"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.442074 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pf44z" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.445168 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-dwv5n" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.498120 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-79p7l"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.512314 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.529416 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ftlx" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.530422 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lpxm\" (UniqueName: \"kubernetes.io/projected/3c945977-c07e-4ef4-9370-d2fa1264eee0-kube-api-access-2lpxm\") pod \"placement-operator-controller-manager-78f8948974-88stb\" (UID: \"3c945977-c07e-4ef4-9370-d2fa1264eee0\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-88stb" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.530669 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtkwl\" (UniqueName: \"kubernetes.io/projected/65429c05-8bff-4545-a085-e4c7faa29fa3-kube-api-access-qtkwl\") pod \"swift-operator-controller-manager-9d58d64bc-pf44z\" (UID: \"65429c05-8bff-4545-a085-e4c7faa29fa3\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pf44z" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.570050 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lpxm\" (UniqueName: \"kubernetes.io/projected/3c945977-c07e-4ef4-9370-d2fa1264eee0-kube-api-access-2lpxm\") pod \"placement-operator-controller-manager-78f8948974-88stb\" (UID: \"3c945977-c07e-4ef4-9370-d2fa1264eee0\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-88stb" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.592063 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-pf44z"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.627859 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-954d6f856-2z6fx"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.629393 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-2z6fx" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.636743 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtkwl\" (UniqueName: \"kubernetes.io/projected/65429c05-8bff-4545-a085-e4c7faa29fa3-kube-api-access-qtkwl\") pod \"swift-operator-controller-manager-9d58d64bc-pf44z\" (UID: \"65429c05-8bff-4545-a085-e4c7faa29fa3\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pf44z" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.636854 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert\") pod \"infra-operator-controller-manager-78d48bff9d-x27kt\" (UID: \"55bb5bf6-cc44-43f9-84e5-1137401fd188\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" Dec 10 19:15:50 crc kubenswrapper[4828]: E1210 19:15:50.636986 4828 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 19:15:50 crc kubenswrapper[4828]: E1210 19:15:50.637038 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert podName:55bb5bf6-cc44-43f9-84e5-1137401fd188 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:51.637017369 +0000 UTC m=+1232.147628374 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert") pod "infra-operator-controller-manager-78d48bff9d-x27kt" (UID: "55bb5bf6-cc44-43f9-84e5-1137401fd188") : secret "infra-operator-webhook-server-cert" not found Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.637397 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-z8rfx" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.656250 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-954d6f856-2z6fx"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.656616 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-j64bx" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.669710 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtkwl\" (UniqueName: \"kubernetes.io/projected/65429c05-8bff-4545-a085-e4c7faa29fa3-kube-api-access-qtkwl\") pod \"swift-operator-controller-manager-9d58d64bc-pf44z\" (UID: \"65429c05-8bff-4545-a085-e4c7faa29fa3\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pf44z" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.739581 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5rxr\" (UniqueName: \"kubernetes.io/projected/e0825966-a2fc-4a99-9fc8-d256f885a2c5-kube-api-access-f5rxr\") pod \"telemetry-operator-controller-manager-954d6f856-2z6fx\" (UID: \"e0825966-a2fc-4a99-9fc8-d256f885a2c5\") " pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-2z6fx" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.748043 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-88stb" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.777415 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-942ph"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.778934 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-942ph" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.784694 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-4nt6t" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.788113 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pf44z" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.793122 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-942ph"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.825943 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.833567 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.837191 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-j5nfd" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.842748 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fk7hbh\" (UID: \"4614b8b1-2082-4b05-8473-404f0f82772b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.842813 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmqsn\" (UniqueName: \"kubernetes.io/projected/d7200137-65c3-48b5-bc24-5cf17caa07f4-kube-api-access-gmqsn\") pod \"test-operator-controller-manager-5854674fcc-942ph\" (UID: \"d7200137-65c3-48b5-bc24-5cf17caa07f4\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-942ph" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.842857 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5rxr\" (UniqueName: \"kubernetes.io/projected/e0825966-a2fc-4a99-9fc8-d256f885a2c5-kube-api-access-f5rxr\") pod \"telemetry-operator-controller-manager-954d6f856-2z6fx\" (UID: \"e0825966-a2fc-4a99-9fc8-d256f885a2c5\") " pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-2z6fx" Dec 10 19:15:50 crc kubenswrapper[4828]: E1210 19:15:50.847279 4828 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:15:50 crc kubenswrapper[4828]: E1210 19:15:50.849917 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert podName:4614b8b1-2082-4b05-8473-404f0f82772b nodeName:}" failed. No retries permitted until 2025-12-10 19:15:51.847386142 +0000 UTC m=+1232.357997147 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fk7hbh" (UID: "4614b8b1-2082-4b05-8473-404f0f82772b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.865625 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5rxr\" (UniqueName: \"kubernetes.io/projected/e0825966-a2fc-4a99-9fc8-d256f885a2c5-kube-api-access-f5rxr\") pod \"telemetry-operator-controller-manager-954d6f856-2z6fx\" (UID: \"e0825966-a2fc-4a99-9fc8-d256f885a2c5\") " pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-2z6fx" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.869966 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.901891 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.904702 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.906755 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.908167 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-jvmmz" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.908252 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.918779 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.947965 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppmrr\" (UniqueName: \"kubernetes.io/projected/1c7fb44a-c32b-49f0-8220-9cb69be63935-kube-api-access-ppmrr\") pod \"watcher-operator-controller-manager-75944c9b7-s84vk\" (UID: \"1c7fb44a-c32b-49f0-8220-9cb69be63935\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.948073 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmqsn\" (UniqueName: \"kubernetes.io/projected/d7200137-65c3-48b5-bc24-5cf17caa07f4-kube-api-access-gmqsn\") pod \"test-operator-controller-manager-5854674fcc-942ph\" (UID: \"d7200137-65c3-48b5-bc24-5cf17caa07f4\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-942ph" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.968061 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cv2bf"] Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.969030 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmqsn\" (UniqueName: \"kubernetes.io/projected/d7200137-65c3-48b5-bc24-5cf17caa07f4-kube-api-access-gmqsn\") pod \"test-operator-controller-manager-5854674fcc-942ph\" (UID: \"d7200137-65c3-48b5-bc24-5cf17caa07f4\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-942ph" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.976261 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cv2bf" Dec 10 19:15:50 crc kubenswrapper[4828]: I1210 19:15:50.979284 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-s79ls" Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.004789 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cv2bf"] Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.013388 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-2z6fx" Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.050453 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tktl6\" (UniqueName: \"kubernetes.io/projected/0a273a13-6393-4a24-85c8-8ac01dbf0e5c-kube-api-access-tktl6\") pod \"rabbitmq-cluster-operator-manager-668c99d594-cv2bf\" (UID: \"0a273a13-6393-4a24-85c8-8ac01dbf0e5c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cv2bf" Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.050485 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.050518 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.050753 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh72g\" (UniqueName: \"kubernetes.io/projected/25aaba8d-1069-4360-a0b4-af30543dbb02-kube-api-access-fh72g\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.050949 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppmrr\" (UniqueName: \"kubernetes.io/projected/1c7fb44a-c32b-49f0-8220-9cb69be63935-kube-api-access-ppmrr\") pod \"watcher-operator-controller-manager-75944c9b7-s84vk\" (UID: \"1c7fb44a-c32b-49f0-8220-9cb69be63935\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk" Dec 10 19:15:51 crc kubenswrapper[4828]: W1210 19:15:51.059322 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbad66884_aa3d_4683_abc6_2387140d407f.slice/crio-4615cf2a52bb145c07b2a0cb1c132114bd09f14b91d14877f46e576ee9f8c110 WatchSource:0}: Error finding container 4615cf2a52bb145c07b2a0cb1c132114bd09f14b91d14877f46e576ee9f8c110: Status 404 returned error can't find the container with id 4615cf2a52bb145c07b2a0cb1c132114bd09f14b91d14877f46e576ee9f8c110 Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.071569 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-vhf9j"] Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.074393 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppmrr\" (UniqueName: \"kubernetes.io/projected/1c7fb44a-c32b-49f0-8220-9cb69be63935-kube-api-access-ppmrr\") pod \"watcher-operator-controller-manager-75944c9b7-s84vk\" (UID: \"1c7fb44a-c32b-49f0-8220-9cb69be63935\") " pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk" Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.086424 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mmqkg"] Dec 10 19:15:51 crc kubenswrapper[4828]: W1210 19:15:51.090396 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0544042_b6b7_4a9d_884a_12859ddea597.slice/crio-6dc6c7de13e1fd6917def35ea95bdeaefef36db36d94064a38163033ae096011 WatchSource:0}: Error finding container 6dc6c7de13e1fd6917def35ea95bdeaefef36db36d94064a38163033ae096011: Status 404 returned error can't find the container with id 6dc6c7de13e1fd6917def35ea95bdeaefef36db36d94064a38163033ae096011 Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.097218 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-pcfhr"] Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.126612 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-942ph" Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.155206 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh72g\" (UniqueName: \"kubernetes.io/projected/25aaba8d-1069-4360-a0b4-af30543dbb02-kube-api-access-fh72g\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.155412 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tktl6\" (UniqueName: \"kubernetes.io/projected/0a273a13-6393-4a24-85c8-8ac01dbf0e5c-kube-api-access-tktl6\") pod \"rabbitmq-cluster-operator-manager-668c99d594-cv2bf\" (UID: \"0a273a13-6393-4a24-85c8-8ac01dbf0e5c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cv2bf" Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.155445 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.155487 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:15:51 crc kubenswrapper[4828]: E1210 19:15:51.155861 4828 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 19:15:51 crc kubenswrapper[4828]: E1210 19:15:51.155909 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs podName:25aaba8d-1069-4360-a0b4-af30543dbb02 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:51.655889586 +0000 UTC m=+1232.166500591 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs") pod "openstack-operator-controller-manager-7b5585cf46-6j82f" (UID: "25aaba8d-1069-4360-a0b4-af30543dbb02") : secret "webhook-server-cert" not found Dec 10 19:15:51 crc kubenswrapper[4828]: E1210 19:15:51.156195 4828 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 19:15:51 crc kubenswrapper[4828]: E1210 19:15:51.156226 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs podName:25aaba8d-1069-4360-a0b4-af30543dbb02 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:51.656218255 +0000 UTC m=+1232.166829260 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs") pod "openstack-operator-controller-manager-7b5585cf46-6j82f" (UID: "25aaba8d-1069-4360-a0b4-af30543dbb02") : secret "metrics-server-cert" not found Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.173111 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh72g\" (UniqueName: \"kubernetes.io/projected/25aaba8d-1069-4360-a0b4-af30543dbb02-kube-api-access-fh72g\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.176189 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk" Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.176702 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tktl6\" (UniqueName: \"kubernetes.io/projected/0a273a13-6393-4a24-85c8-8ac01dbf0e5c-kube-api-access-tktl6\") pod \"rabbitmq-cluster-operator-manager-668c99d594-cv2bf\" (UID: \"0a273a13-6393-4a24-85c8-8ac01dbf0e5c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cv2bf" Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.196193 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cv2bf" Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.488936 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-79p7l"] Dec 10 19:15:51 crc kubenswrapper[4828]: W1210 19:15:51.490580 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe915c9b_8dfe_4aab_8e74_5375e365e623.slice/crio-05b123dadbfaeafc7dc802a7e3f2d970e6c43cb3862d62bf05943b3c8551e7a9 WatchSource:0}: Error finding container 05b123dadbfaeafc7dc802a7e3f2d970e6c43cb3862d62bf05943b3c8551e7a9: Status 404 returned error can't find the container with id 05b123dadbfaeafc7dc802a7e3f2d970e6c43cb3862d62bf05943b3c8551e7a9 Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.494790 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6bqjc"] Dec 10 19:15:51 crc kubenswrapper[4828]: W1210 19:15:51.501394 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85a6c448_c39f_4ec5_b1f7_4f7ff1bc43a0.slice/crio-98db874526de5f1879622d3a020bb7fb60587d0bd52739e3ee93e6b6cbcb68fe WatchSource:0}: Error finding container 98db874526de5f1879622d3a020bb7fb60587d0bd52739e3ee93e6b6cbcb68fe: Status 404 returned error can't find the container with id 98db874526de5f1879622d3a020bb7fb60587d0bd52739e3ee93e6b6cbcb68fe Dec 10 19:15:51 crc kubenswrapper[4828]: W1210 19:15:51.502201 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23431602_592f_4c26_a5ab_1ada81fda48a.slice/crio-779a27e615aae9a96434e1da349bb0643601aa6753d3531652f639feaf1b3a05 WatchSource:0}: Error finding container 779a27e615aae9a96434e1da349bb0643601aa6753d3531652f639feaf1b3a05: Status 404 returned error can't find the container with id 779a27e615aae9a96434e1da349bb0643601aa6753d3531652f639feaf1b3a05 Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.505301 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-f7cn8"] Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.664583 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert\") pod \"infra-operator-controller-manager-78d48bff9d-x27kt\" (UID: \"55bb5bf6-cc44-43f9-84e5-1137401fd188\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.664655 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.664691 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:15:51 crc kubenswrapper[4828]: E1210 19:15:51.664855 4828 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 19:15:51 crc kubenswrapper[4828]: E1210 19:15:51.664886 4828 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 19:15:51 crc kubenswrapper[4828]: E1210 19:15:51.664923 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs podName:25aaba8d-1069-4360-a0b4-af30543dbb02 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:52.664904684 +0000 UTC m=+1233.175515689 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs") pod "openstack-operator-controller-manager-7b5585cf46-6j82f" (UID: "25aaba8d-1069-4360-a0b4-af30543dbb02") : secret "metrics-server-cert" not found Dec 10 19:15:51 crc kubenswrapper[4828]: E1210 19:15:51.664951 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs podName:25aaba8d-1069-4360-a0b4-af30543dbb02 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:52.664930695 +0000 UTC m=+1233.175541780 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs") pod "openstack-operator-controller-manager-7b5585cf46-6j82f" (UID: "25aaba8d-1069-4360-a0b4-af30543dbb02") : secret "webhook-server-cert" not found Dec 10 19:15:51 crc kubenswrapper[4828]: E1210 19:15:51.665266 4828 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 19:15:51 crc kubenswrapper[4828]: E1210 19:15:51.665295 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert podName:55bb5bf6-cc44-43f9-84e5-1137401fd188 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:53.665287264 +0000 UTC m=+1234.175898269 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert") pod "infra-operator-controller-manager-78d48bff9d-x27kt" (UID: "55bb5bf6-cc44-43f9-84e5-1137401fd188") : secret "infra-operator-webhook-server-cert" not found Dec 10 19:15:51 crc kubenswrapper[4828]: W1210 19:15:51.677989 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f635d23_992f_4a05_9896_c38d41ec853d.slice/crio-a3bb87cb3d21d7592753c50ef3120f3549d20dd24d7673f5b27abc3b26794a29 WatchSource:0}: Error finding container a3bb87cb3d21d7592753c50ef3120f3549d20dd24d7673f5b27abc3b26794a29: Status 404 returned error can't find the container with id a3bb87cb3d21d7592753c50ef3120f3549d20dd24d7673f5b27abc3b26794a29 Dec 10 19:15:51 crc kubenswrapper[4828]: W1210 19:15:51.680464 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8991af61_9866_4ae9_b7a7_7545833c4485.slice/crio-296e4db2b9471e376cde970b47a59500fa586db64fa5be5315f13f08870479da WatchSource:0}: Error finding container 296e4db2b9471e376cde970b47a59500fa586db64fa5be5315f13f08870479da: Status 404 returned error can't find the container with id 296e4db2b9471e376cde970b47a59500fa586db64fa5be5315f13f08870479da Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.683208 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-cfhjs"] Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.706647 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-njhhp"] Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.721350 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-dxcl8"] Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.732686 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-wtcz4"] Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.868309 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fk7hbh\" (UID: \"4614b8b1-2082-4b05-8473-404f0f82772b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" Dec 10 19:15:51 crc kubenswrapper[4828]: E1210 19:15:51.869318 4828 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:15:51 crc kubenswrapper[4828]: E1210 19:15:51.869386 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert podName:4614b8b1-2082-4b05-8473-404f0f82772b nodeName:}" failed. No retries permitted until 2025-12-10 19:15:53.869366802 +0000 UTC m=+1234.379977857 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fk7hbh" (UID: "4614b8b1-2082-4b05-8473-404f0f82772b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.950567 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-dxcl8" event={"ID":"4246d7f6-cfb7-4d1d-aac0-6ba9db04a50b","Type":"ContainerStarted","Data":"482041c440c04de5b6e48ac431aa72d6537ef52ff0011633fe06e76c20126f50"} Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.952209 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-79p7l" event={"ID":"be915c9b-8dfe-4aab-8e74-5375e365e623","Type":"ContainerStarted","Data":"05b123dadbfaeafc7dc802a7e3f2d970e6c43cb3862d62bf05943b3c8551e7a9"} Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.953474 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-njhhp" event={"ID":"8f635d23-992f-4a05-9896-c38d41ec853d","Type":"ContainerStarted","Data":"a3bb87cb3d21d7592753c50ef3120f3549d20dd24d7673f5b27abc3b26794a29"} Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.954949 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6bqjc" event={"ID":"23431602-592f-4c26-a5ab-1ada81fda48a","Type":"ContainerStarted","Data":"779a27e615aae9a96434e1da349bb0643601aa6753d3531652f639feaf1b3a05"} Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.956322 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-cfhjs" event={"ID":"8991af61-9866-4ae9-b7a7-7545833c4485","Type":"ContainerStarted","Data":"296e4db2b9471e376cde970b47a59500fa586db64fa5be5315f13f08870479da"} Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.957469 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mmqkg" event={"ID":"bad66884-aa3d-4683-abc6-2387140d407f","Type":"ContainerStarted","Data":"4615cf2a52bb145c07b2a0cb1c132114bd09f14b91d14877f46e576ee9f8c110"} Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.958502 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vhf9j" event={"ID":"1f3d156d-ddc6-47a2-9c40-5b675e535b37","Type":"ContainerStarted","Data":"2f1ff184d1545fceb815c698f7e44a00909f773bb8fa217499a6563648f44ae1"} Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.959531 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-wtcz4" event={"ID":"bd786401-638a-4830-8a28-cc01b03d311d","Type":"ContainerStarted","Data":"58cbff9cc51c39c30f028ae5c73bbed3a1bff13b26028eca859cd004bc5fd4f5"} Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.960632 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-f7cn8" event={"ID":"85a6c448-c39f-4ec5-b1f7-4f7ff1bc43a0","Type":"ContainerStarted","Data":"98db874526de5f1879622d3a020bb7fb60587d0bd52739e3ee93e6b6cbcb68fe"} Dec 10 19:15:51 crc kubenswrapper[4828]: I1210 19:15:51.967623 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-pcfhr" event={"ID":"b0544042-b6b7-4a9d-884a-12859ddea597","Type":"ContainerStarted","Data":"6dc6c7de13e1fd6917def35ea95bdeaefef36db36d94064a38163033ae096011"} Dec 10 19:15:52 crc kubenswrapper[4828]: I1210 19:15:52.393152 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-942ph"] Dec 10 19:15:52 crc kubenswrapper[4828]: I1210 19:15:52.405878 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-pf44z"] Dec 10 19:15:52 crc kubenswrapper[4828]: I1210 19:15:52.438355 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ftlx"] Dec 10 19:15:52 crc kubenswrapper[4828]: I1210 19:15:52.478655 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-954d6f856-2z6fx"] Dec 10 19:15:52 crc kubenswrapper[4828]: I1210 19:15:52.498045 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-smtjq"] Dec 10 19:15:52 crc kubenswrapper[4828]: W1210 19:15:52.487998 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7200137_65c3_48b5_bc24_5cf17caa07f4.slice/crio-6b8a5914cebad7fcd462768f8f6445fbd74e9dd0e3d603f73b0ffd4377d53d17 WatchSource:0}: Error finding container 6b8a5914cebad7fcd462768f8f6445fbd74e9dd0e3d603f73b0ffd4377d53d17: Status 404 returned error can't find the container with id 6b8a5914cebad7fcd462768f8f6445fbd74e9dd0e3d603f73b0ffd4377d53d17 Dec 10 19:15:52 crc kubenswrapper[4828]: W1210 19:15:52.501751 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c945977_c07e_4ef4_9370_d2fa1264eee0.slice/crio-5eec449e98989126a322ba242af6b31ea9a96a13845165868cdf0080da8ffdcb WatchSource:0}: Error finding container 5eec449e98989126a322ba242af6b31ea9a96a13845165868cdf0080da8ffdcb: Status 404 returned error can't find the container with id 5eec449e98989126a322ba242af6b31ea9a96a13845165868cdf0080da8ffdcb Dec 10 19:15:52 crc kubenswrapper[4828]: I1210 19:15:52.517313 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cv2bf"] Dec 10 19:15:52 crc kubenswrapper[4828]: I1210 19:15:52.528880 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-88stb"] Dec 10 19:15:52 crc kubenswrapper[4828]: I1210 19:15:52.540643 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-j64bx"] Dec 10 19:15:52 crc kubenswrapper[4828]: I1210 19:15:52.558526 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg"] Dec 10 19:15:52 crc kubenswrapper[4828]: E1210 19:15:52.558538 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2lpxm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-88stb_openstack-operators(3c945977-c07e-4ef4-9370-d2fa1264eee0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:15:52 crc kubenswrapper[4828]: E1210 19:15:52.561610 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2lpxm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-88stb_openstack-operators(3c945977-c07e-4ef4-9370-d2fa1264eee0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:15:52 crc kubenswrapper[4828]: I1210 19:15:52.562289 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk"] Dec 10 19:15:52 crc kubenswrapper[4828]: E1210 19:15:52.563977 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-88stb" podUID="3c945977-c07e-4ef4-9370-d2fa1264eee0" Dec 10 19:15:52 crc kubenswrapper[4828]: W1210 19:15:52.615487 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c7fb44a_c32b_49f0_8220_9cb69be63935.slice/crio-9ba19a2eb4af8fab95d94cc0fac775d7e6afa4b29777b3cf53e4220f951bb232 WatchSource:0}: Error finding container 9ba19a2eb4af8fab95d94cc0fac775d7e6afa4b29777b3cf53e4220f951bb232: Status 404 returned error can't find the container with id 9ba19a2eb4af8fab95d94cc0fac775d7e6afa4b29777b3cf53e4220f951bb232 Dec 10 19:15:52 crc kubenswrapper[4828]: E1210 19:15:52.615857 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bwmjr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-s5jjg_openstack-operators(e87fb7c5-676f-4549-82e3-2696306f651b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:15:52 crc kubenswrapper[4828]: E1210 19:15:52.621521 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bwmjr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-s5jjg_openstack-operators(e87fb7c5-676f-4549-82e3-2696306f651b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:15:52 crc kubenswrapper[4828]: E1210 19:15:52.622871 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg" podUID="e87fb7c5-676f-4549-82e3-2696306f651b" Dec 10 19:15:52 crc kubenswrapper[4828]: E1210 19:15:52.635096 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ppmrr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-75944c9b7-s84vk_openstack-operators(1c7fb44a-c32b-49f0-8220-9cb69be63935): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:15:52 crc kubenswrapper[4828]: E1210 19:15:52.645015 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ppmrr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-75944c9b7-s84vk_openstack-operators(1c7fb44a-c32b-49f0-8220-9cb69be63935): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 10 19:15:52 crc kubenswrapper[4828]: E1210 19:15:52.658564 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk" podUID="1c7fb44a-c32b-49f0-8220-9cb69be63935" Dec 10 19:15:52 crc kubenswrapper[4828]: I1210 19:15:52.687905 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:15:52 crc kubenswrapper[4828]: I1210 19:15:52.687958 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:15:52 crc kubenswrapper[4828]: E1210 19:15:52.688117 4828 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 19:15:52 crc kubenswrapper[4828]: E1210 19:15:52.688164 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs podName:25aaba8d-1069-4360-a0b4-af30543dbb02 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:54.688149798 +0000 UTC m=+1235.198760803 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs") pod "openstack-operator-controller-manager-7b5585cf46-6j82f" (UID: "25aaba8d-1069-4360-a0b4-af30543dbb02") : secret "webhook-server-cert" not found Dec 10 19:15:52 crc kubenswrapper[4828]: E1210 19:15:52.688206 4828 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 19:15:52 crc kubenswrapper[4828]: E1210 19:15:52.688259 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs podName:25aaba8d-1069-4360-a0b4-af30543dbb02 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:54.6882188 +0000 UTC m=+1235.198829805 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs") pod "openstack-operator-controller-manager-7b5585cf46-6j82f" (UID: "25aaba8d-1069-4360-a0b4-af30543dbb02") : secret "metrics-server-cert" not found Dec 10 19:15:52 crc kubenswrapper[4828]: I1210 19:15:52.992484 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-j64bx" event={"ID":"70814aec-fe59-4dc6-a085-c5a5da7b93a5","Type":"ContainerStarted","Data":"fb7df35c30931b63376bcd7a2168b0e763f0db18e10bf18b7750b48cf0d2878b"} Dec 10 19:15:53 crc kubenswrapper[4828]: I1210 19:15:52.995984 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg" event={"ID":"e87fb7c5-676f-4549-82e3-2696306f651b","Type":"ContainerStarted","Data":"9af2a56c753a552b35bf2b3a22f1852356edab81f6514feb8864489213cbece8"} Dec 10 19:15:53 crc kubenswrapper[4828]: E1210 19:15:52.998819 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg" podUID="e87fb7c5-676f-4549-82e3-2696306f651b" Dec 10 19:15:53 crc kubenswrapper[4828]: I1210 19:15:53.002791 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-2z6fx" event={"ID":"e0825966-a2fc-4a99-9fc8-d256f885a2c5","Type":"ContainerStarted","Data":"944e504f2b4483828647cb62c13068df62f07e95903e1cc78796ba972a53a073"} Dec 10 19:15:53 crc kubenswrapper[4828]: I1210 19:15:53.008864 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-smtjq" event={"ID":"ab0dbc03-b62b-40c9-98b9-7bb192cddff3","Type":"ContainerStarted","Data":"93e42b24f0c8de869c5d9643b6f6f44a1cd1b98086cabe010f96a04e2e484812"} Dec 10 19:15:53 crc kubenswrapper[4828]: I1210 19:15:53.013730 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk" event={"ID":"1c7fb44a-c32b-49f0-8220-9cb69be63935","Type":"ContainerStarted","Data":"9ba19a2eb4af8fab95d94cc0fac775d7e6afa4b29777b3cf53e4220f951bb232"} Dec 10 19:15:53 crc kubenswrapper[4828]: I1210 19:15:53.021772 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-942ph" event={"ID":"d7200137-65c3-48b5-bc24-5cf17caa07f4","Type":"ContainerStarted","Data":"6b8a5914cebad7fcd462768f8f6445fbd74e9dd0e3d603f73b0ffd4377d53d17"} Dec 10 19:15:53 crc kubenswrapper[4828]: I1210 19:15:53.024262 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-88stb" event={"ID":"3c945977-c07e-4ef4-9370-d2fa1264eee0","Type":"ContainerStarted","Data":"5eec449e98989126a322ba242af6b31ea9a96a13845165868cdf0080da8ffdcb"} Dec 10 19:15:53 crc kubenswrapper[4828]: E1210 19:15:53.024790 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk" podUID="1c7fb44a-c32b-49f0-8220-9cb69be63935" Dec 10 19:15:53 crc kubenswrapper[4828]: I1210 19:15:53.025771 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ftlx" event={"ID":"be987ffc-f75a-4364-ac29-37fbc9bb7200","Type":"ContainerStarted","Data":"f11bfe04b52b998ccadeb6fccc553133a3bee82bc41daf58bfed2d7c9291eb6f"} Dec 10 19:15:53 crc kubenswrapper[4828]: E1210 19:15:53.026173 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-88stb" podUID="3c945977-c07e-4ef4-9370-d2fa1264eee0" Dec 10 19:15:53 crc kubenswrapper[4828]: I1210 19:15:53.028043 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cv2bf" event={"ID":"0a273a13-6393-4a24-85c8-8ac01dbf0e5c","Type":"ContainerStarted","Data":"2b5a00fbf0ff49c76146b0558166cd6e85106da29e92a78bc44fbb7ad99f3cd2"} Dec 10 19:15:53 crc kubenswrapper[4828]: I1210 19:15:53.034700 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pf44z" event={"ID":"65429c05-8bff-4545-a085-e4c7faa29fa3","Type":"ContainerStarted","Data":"28e504fd911dc03f9b5bbca14d2f93aeddd01c6a1942fd8e0b642cb9477f6627"} Dec 10 19:15:53 crc kubenswrapper[4828]: I1210 19:15:53.733072 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert\") pod \"infra-operator-controller-manager-78d48bff9d-x27kt\" (UID: \"55bb5bf6-cc44-43f9-84e5-1137401fd188\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" Dec 10 19:15:53 crc kubenswrapper[4828]: E1210 19:15:53.733234 4828 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 19:15:53 crc kubenswrapper[4828]: E1210 19:15:53.733525 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert podName:55bb5bf6-cc44-43f9-84e5-1137401fd188 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:57.733507132 +0000 UTC m=+1238.244118137 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert") pod "infra-operator-controller-manager-78d48bff9d-x27kt" (UID: "55bb5bf6-cc44-43f9-84e5-1137401fd188") : secret "infra-operator-webhook-server-cert" not found Dec 10 19:15:53 crc kubenswrapper[4828]: I1210 19:15:53.937166 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fk7hbh\" (UID: \"4614b8b1-2082-4b05-8473-404f0f82772b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" Dec 10 19:15:53 crc kubenswrapper[4828]: E1210 19:15:53.937410 4828 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:15:53 crc kubenswrapper[4828]: E1210 19:15:53.937472 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert podName:4614b8b1-2082-4b05-8473-404f0f82772b nodeName:}" failed. No retries permitted until 2025-12-10 19:15:57.937454546 +0000 UTC m=+1238.448065551 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fk7hbh" (UID: "4614b8b1-2082-4b05-8473-404f0f82772b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:15:54 crc kubenswrapper[4828]: E1210 19:15:54.120929 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-88stb" podUID="3c945977-c07e-4ef4-9370-d2fa1264eee0" Dec 10 19:15:54 crc kubenswrapper[4828]: E1210 19:15:54.121029 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg" podUID="e87fb7c5-676f-4549-82e3-2696306f651b" Dec 10 19:15:54 crc kubenswrapper[4828]: E1210 19:15:54.121086 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk" podUID="1c7fb44a-c32b-49f0-8220-9cb69be63935" Dec 10 19:15:54 crc kubenswrapper[4828]: I1210 19:15:54.768036 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:15:54 crc kubenswrapper[4828]: I1210 19:15:54.768098 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:15:54 crc kubenswrapper[4828]: E1210 19:15:54.768180 4828 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 19:15:54 crc kubenswrapper[4828]: E1210 19:15:54.768241 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs podName:25aaba8d-1069-4360-a0b4-af30543dbb02 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:58.768222136 +0000 UTC m=+1239.278833141 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs") pod "openstack-operator-controller-manager-7b5585cf46-6j82f" (UID: "25aaba8d-1069-4360-a0b4-af30543dbb02") : secret "metrics-server-cert" not found Dec 10 19:15:54 crc kubenswrapper[4828]: E1210 19:15:54.768289 4828 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 19:15:54 crc kubenswrapper[4828]: E1210 19:15:54.768317 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs podName:25aaba8d-1069-4360-a0b4-af30543dbb02 nodeName:}" failed. No retries permitted until 2025-12-10 19:15:58.768307589 +0000 UTC m=+1239.278918594 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs") pod "openstack-operator-controller-manager-7b5585cf46-6j82f" (UID: "25aaba8d-1069-4360-a0b4-af30543dbb02") : secret "webhook-server-cert" not found Dec 10 19:15:57 crc kubenswrapper[4828]: I1210 19:15:57.734460 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert\") pod \"infra-operator-controller-manager-78d48bff9d-x27kt\" (UID: \"55bb5bf6-cc44-43f9-84e5-1137401fd188\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" Dec 10 19:15:57 crc kubenswrapper[4828]: E1210 19:15:57.734578 4828 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 10 19:15:57 crc kubenswrapper[4828]: E1210 19:15:57.735234 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert podName:55bb5bf6-cc44-43f9-84e5-1137401fd188 nodeName:}" failed. No retries permitted until 2025-12-10 19:16:05.735210945 +0000 UTC m=+1246.245821950 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert") pod "infra-operator-controller-manager-78d48bff9d-x27kt" (UID: "55bb5bf6-cc44-43f9-84e5-1137401fd188") : secret "infra-operator-webhook-server-cert" not found Dec 10 19:15:57 crc kubenswrapper[4828]: I1210 19:15:57.940220 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fk7hbh\" (UID: \"4614b8b1-2082-4b05-8473-404f0f82772b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" Dec 10 19:15:57 crc kubenswrapper[4828]: E1210 19:15:57.940488 4828 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:15:57 crc kubenswrapper[4828]: E1210 19:15:57.940589 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert podName:4614b8b1-2082-4b05-8473-404f0f82772b nodeName:}" failed. No retries permitted until 2025-12-10 19:16:05.940565187 +0000 UTC m=+1246.451176262 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fk7hbh" (UID: "4614b8b1-2082-4b05-8473-404f0f82772b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 10 19:15:58 crc kubenswrapper[4828]: I1210 19:15:58.856975 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:15:58 crc kubenswrapper[4828]: I1210 19:15:58.857052 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:15:58 crc kubenswrapper[4828]: E1210 19:15:58.857138 4828 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 10 19:15:58 crc kubenswrapper[4828]: E1210 19:15:58.857211 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs podName:25aaba8d-1069-4360-a0b4-af30543dbb02 nodeName:}" failed. No retries permitted until 2025-12-10 19:16:06.857194616 +0000 UTC m=+1247.367805621 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs") pod "openstack-operator-controller-manager-7b5585cf46-6j82f" (UID: "25aaba8d-1069-4360-a0b4-af30543dbb02") : secret "metrics-server-cert" not found Dec 10 19:15:58 crc kubenswrapper[4828]: E1210 19:15:58.857283 4828 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 10 19:15:58 crc kubenswrapper[4828]: E1210 19:15:58.857373 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs podName:25aaba8d-1069-4360-a0b4-af30543dbb02 nodeName:}" failed. No retries permitted until 2025-12-10 19:16:06.85734529 +0000 UTC m=+1247.367956325 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs") pod "openstack-operator-controller-manager-7b5585cf46-6j82f" (UID: "25aaba8d-1069-4360-a0b4-af30543dbb02") : secret "webhook-server-cert" not found Dec 10 19:16:05 crc kubenswrapper[4828]: I1210 19:16:05.787138 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert\") pod \"infra-operator-controller-manager-78d48bff9d-x27kt\" (UID: \"55bb5bf6-cc44-43f9-84e5-1137401fd188\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" Dec 10 19:16:05 crc kubenswrapper[4828]: I1210 19:16:05.797222 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55bb5bf6-cc44-43f9-84e5-1137401fd188-cert\") pod \"infra-operator-controller-manager-78d48bff9d-x27kt\" (UID: \"55bb5bf6-cc44-43f9-84e5-1137401fd188\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" Dec 10 19:16:05 crc kubenswrapper[4828]: I1210 19:16:05.990109 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fk7hbh\" (UID: \"4614b8b1-2082-4b05-8473-404f0f82772b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" Dec 10 19:16:05 crc kubenswrapper[4828]: I1210 19:16:05.993682 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4614b8b1-2082-4b05-8473-404f0f82772b-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fk7hbh\" (UID: \"4614b8b1-2082-4b05-8473-404f0f82772b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" Dec 10 19:16:06 crc kubenswrapper[4828]: I1210 19:16:06.016530 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" Dec 10 19:16:06 crc kubenswrapper[4828]: I1210 19:16:06.028111 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" Dec 10 19:16:06 crc kubenswrapper[4828]: I1210 19:16:06.906042 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:16:06 crc kubenswrapper[4828]: I1210 19:16:06.906450 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:16:06 crc kubenswrapper[4828]: I1210 19:16:06.910491 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-metrics-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:16:06 crc kubenswrapper[4828]: I1210 19:16:06.911140 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/25aaba8d-1069-4360-a0b4-af30543dbb02-webhook-certs\") pod \"openstack-operator-controller-manager-7b5585cf46-6j82f\" (UID: \"25aaba8d-1069-4360-a0b4-af30543dbb02\") " pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:16:07 crc kubenswrapper[4828]: I1210 19:16:07.123492 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:16:18 crc kubenswrapper[4828]: E1210 19:16:18.327385 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad" Dec 10 19:16:18 crc kubenswrapper[4828]: E1210 19:16:18.328250 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:424da951f13f1fbe9083215dc9f5088f90676dd813f01fdf3c1a8639b61cbaad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7q6sn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-6bqjc_openstack-operators(23431602-592f-4c26-a5ab-1ada81fda48a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:19 crc kubenswrapper[4828]: E1210 19:16:19.186214 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027" Dec 10 19:16:19 crc kubenswrapper[4828]: E1210 19:16:19.186727 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8m9pq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-5697bb5779-vhf9j_openstack-operators(1f3d156d-ddc6-47a2-9c40-5b675e535b37): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:24 crc kubenswrapper[4828]: E1210 19:16:24.168396 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87" Dec 10 19:16:24 crc kubenswrapper[4828]: E1210 19:16:24.169170 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zw6cg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-967d97867-f7cn8_openstack-operators(85a6c448-c39f-4ec5-b1f7-4f7ff1bc43a0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:27 crc kubenswrapper[4828]: E1210 19:16:27.185350 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 10 19:16:27 crc kubenswrapper[4828]: E1210 19:16:27.185586 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s5rd2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-mmqkg_openstack-operators(bad66884-aa3d-4683-abc6-2387140d407f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:28 crc kubenswrapper[4828]: E1210 19:16:28.472757 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 10 19:16:28 crc kubenswrapper[4828]: E1210 19:16:28.473190 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nt46s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-dxcl8_openstack-operators(4246d7f6-cfb7-4d1d-aac0-6ba9db04a50b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:29 crc kubenswrapper[4828]: E1210 19:16:29.120965 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a" Dec 10 19:16:29 crc kubenswrapper[4828]: E1210 19:16:29.121167 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6w428,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-697fb699cf-wtcz4_openstack-operators(bd786401-638a-4830-8a28-cc01b03d311d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:34 crc kubenswrapper[4828]: E1210 19:16:34.173490 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a" Dec 10 19:16:34 crc kubenswrapper[4828]: E1210 19:16:34.174316 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kwp2h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-njhhp_openstack-operators(8f635d23-992f-4a05-9896-c38d41ec853d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:46 crc kubenswrapper[4828]: E1210 19:16:46.570999 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea" Dec 10 19:16:46 crc kubenswrapper[4828]: E1210 19:16:46.571735 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d7c98,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-pcfhr_openstack-operators(b0544042-b6b7-4a9d-884a-12859ddea597): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:50 crc kubenswrapper[4828]: E1210 19:16:50.989923 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3" Dec 10 19:16:50 crc kubenswrapper[4828]: E1210 19:16:50.990678 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:981b6a8f95934a86c5f10ef6e198b07265aeba7f11cf84b9ccd13dfaf06f3ca3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lrxgf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-6c677c69b-cfhjs_openstack-operators(8991af61-9866-4ae9-b7a7-7545833c4485): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:51 crc kubenswrapper[4828]: I1210 19:16:51.230889 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:16:51 crc kubenswrapper[4828]: I1210 19:16:51.230952 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:16:51 crc kubenswrapper[4828]: E1210 19:16:51.851141 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.65:5001/openstack-k8s-operators/telemetry-operator:c4794e7165126ca78a1af546bb4ba50c90b5c4e1" Dec 10 19:16:51 crc kubenswrapper[4828]: E1210 19:16:51.851200 4828 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.65:5001/openstack-k8s-operators/telemetry-operator:c4794e7165126ca78a1af546bb4ba50c90b5c4e1" Dec 10 19:16:51 crc kubenswrapper[4828]: E1210 19:16:51.851343 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.65:5001/openstack-k8s-operators/telemetry-operator:c4794e7165126ca78a1af546bb4ba50c90b5c4e1,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f5rxr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-954d6f856-2z6fx_openstack-operators(e0825966-a2fc-4a99-9fc8-d256f885a2c5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:52 crc kubenswrapper[4828]: E1210 19:16:52.913775 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 10 19:16:52 crc kubenswrapper[4828]: E1210 19:16:52.914268 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2lpxm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-88stb_openstack-operators(3c945977-c07e-4ef4-9370-d2fa1264eee0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:53 crc kubenswrapper[4828]: E1210 19:16:53.716786 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 10 19:16:53 crc kubenswrapper[4828]: E1210 19:16:53.717282 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bwmjr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-s5jjg_openstack-operators(e87fb7c5-676f-4549-82e3-2696306f651b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:54 crc kubenswrapper[4828]: E1210 19:16:54.150236 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 10 19:16:54 crc kubenswrapper[4828]: E1210 19:16:54.150397 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nt46s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-dxcl8_openstack-operators(4246d7f6-cfb7-4d1d-aac0-6ba9db04a50b): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 10 19:16:54 crc kubenswrapper[4828]: E1210 19:16:54.152398 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-dxcl8" podUID="4246d7f6-cfb7-4d1d-aac0-6ba9db04a50b" Dec 10 19:16:54 crc kubenswrapper[4828]: E1210 19:16:54.157739 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a" Dec 10 19:16:54 crc kubenswrapper[4828]: E1210 19:16:54.157952 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ppmrr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-75944c9b7-s84vk_openstack-operators(1c7fb44a-c32b-49f0-8220-9cb69be63935): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:54 crc kubenswrapper[4828]: E1210 19:16:54.168891 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 10 19:16:54 crc kubenswrapper[4828]: E1210 19:16:54.169073 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s5rd2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-mmqkg_openstack-operators(bad66884-aa3d-4683-abc6-2387140d407f): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 10 19:16:54 crc kubenswrapper[4828]: E1210 19:16:54.170266 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mmqkg" podUID="bad66884-aa3d-4683-abc6-2387140d407f" Dec 10 19:16:54 crc kubenswrapper[4828]: E1210 19:16:54.589705 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 10 19:16:54 crc kubenswrapper[4828]: E1210 19:16:54.589902 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nfd26,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-4ftlx_openstack-operators(be987ffc-f75a-4364-ac29-37fbc9bb7200): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:54 crc kubenswrapper[4828]: E1210 19:16:54.997735 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 10 19:16:54 crc kubenswrapper[4828]: E1210 19:16:54.998237 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tktl6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-cv2bf_openstack-operators(0a273a13-6393-4a24-85c8-8ac01dbf0e5c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:54 crc kubenswrapper[4828]: E1210 19:16:54.999457 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cv2bf" podUID="0a273a13-6393-4a24-85c8-8ac01dbf0e5c" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.410733 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.410947 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tqw2j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-j64bx_openstack-operators(70814aec-fe59-4dc6-a085-c5a5da7b93a5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.517567 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.518011 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6w428,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-697fb699cf-wtcz4_openstack-operators(bd786401-638a-4830-8a28-cc01b03d311d): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.519411 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.519507 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kwp2h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-njhhp_openstack-operators(8f635d23-992f-4a05-9896-c38d41ec853d): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.519527 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-wtcz4" podUID="bd786401-638a-4830-8a28-cc01b03d311d" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.520660 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-njhhp" podUID="8f635d23-992f-4a05-9896-c38d41ec853d" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.525439 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.525582 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7q6sn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-79c8c4686c-6bqjc_openstack-operators(23431602-592f-4c26-a5ab-1ada81fda48a): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.526777 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6bqjc" podUID="23431602-592f-4c26-a5ab-1ada81fda48a" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.627662 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cv2bf" podUID="0a273a13-6393-4a24-85c8-8ac01dbf0e5c" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.686719 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.686902 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8m9pq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-5697bb5779-vhf9j_openstack-operators(1f3d156d-ddc6-47a2-9c40-5b675e535b37): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.693166 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vhf9j" podUID="1f3d156d-ddc6-47a2-9c40-5b675e535b37" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.725995 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.726164 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zw6cg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-967d97867-f7cn8_openstack-operators(85a6c448-c39f-4ec5-b1f7-4f7ff1bc43a0): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 10 19:16:55 crc kubenswrapper[4828]: E1210 19:16:55.727431 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/ironic-operator-controller-manager-967d97867-f7cn8" podUID="85a6c448-c39f-4ec5-b1f7-4f7ff1bc43a0" Dec 10 19:16:55 crc kubenswrapper[4828]: I1210 19:16:55.820322 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh"] Dec 10 19:16:55 crc kubenswrapper[4828]: I1210 19:16:55.989610 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f"] Dec 10 19:16:56 crc kubenswrapper[4828]: I1210 19:16:56.021707 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt"] Dec 10 19:17:00 crc kubenswrapper[4828]: W1210 19:17:00.955895 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55bb5bf6_cc44_43f9_84e5_1137401fd188.slice/crio-1a5f93d7f9e6318e7ebb813b2ca9e0ee24315c3c717738c8a8a5c00cf4d813ad WatchSource:0}: Error finding container 1a5f93d7f9e6318e7ebb813b2ca9e0ee24315c3c717738c8a8a5c00cf4d813ad: Status 404 returned error can't find the container with id 1a5f93d7f9e6318e7ebb813b2ca9e0ee24315c3c717738c8a8a5c00cf4d813ad Dec 10 19:17:00 crc kubenswrapper[4828]: W1210 19:17:00.957703 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4614b8b1_2082_4b05_8473_404f0f82772b.slice/crio-d91daaa15ade0056af4348d60cead0d2df1a29df747cc6fa824cdebe994d2382 WatchSource:0}: Error finding container d91daaa15ade0056af4348d60cead0d2df1a29df747cc6fa824cdebe994d2382: Status 404 returned error can't find the container with id d91daaa15ade0056af4348d60cead0d2df1a29df747cc6fa824cdebe994d2382 Dec 10 19:17:01 crc kubenswrapper[4828]: I1210 19:17:01.686551 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-smtjq" event={"ID":"ab0dbc03-b62b-40c9-98b9-7bb192cddff3","Type":"ContainerStarted","Data":"5d2b528294aca9e32ca8775b323a3b0e3793f94c9103615d9b21073a1ecde5a2"} Dec 10 19:17:01 crc kubenswrapper[4828]: I1210 19:17:01.688285 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pf44z" event={"ID":"65429c05-8bff-4545-a085-e4c7faa29fa3","Type":"ContainerStarted","Data":"041aace9b93dee206d29a9a788fdcd7fab2caca6a18f1b0284d34c50d9c74b4e"} Dec 10 19:17:01 crc kubenswrapper[4828]: I1210 19:17:01.689616 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" event={"ID":"55bb5bf6-cc44-43f9-84e5-1137401fd188","Type":"ContainerStarted","Data":"1a5f93d7f9e6318e7ebb813b2ca9e0ee24315c3c717738c8a8a5c00cf4d813ad"} Dec 10 19:17:01 crc kubenswrapper[4828]: I1210 19:17:01.703822 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" event={"ID":"4614b8b1-2082-4b05-8473-404f0f82772b","Type":"ContainerStarted","Data":"d91daaa15ade0056af4348d60cead0d2df1a29df747cc6fa824cdebe994d2382"} Dec 10 19:17:01 crc kubenswrapper[4828]: I1210 19:17:01.712451 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-942ph" event={"ID":"d7200137-65c3-48b5-bc24-5cf17caa07f4","Type":"ContainerStarted","Data":"8c3065b2866b545ad2261767515e89eeb6abbfa1b4484596b083b94330a223bb"} Dec 10 19:17:01 crc kubenswrapper[4828]: I1210 19:17:01.714116 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" event={"ID":"25aaba8d-1069-4360-a0b4-af30543dbb02","Type":"ContainerStarted","Data":"54d8c7143ed7386b4d46b59fa571d2fb3b1643df3c2f378538d99067a5f43ffc"} Dec 10 19:17:01 crc kubenswrapper[4828]: I1210 19:17:01.715630 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-79p7l" event={"ID":"be915c9b-8dfe-4aab-8e74-5375e365e623","Type":"ContainerStarted","Data":"96adf0c05ede8fc9cac335c7701994b8313135668dcec49f3dfc85be5f28e917"} Dec 10 19:17:03 crc kubenswrapper[4828]: E1210 19:17:03.212896 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-dxcl8" podUID="4246d7f6-cfb7-4d1d-aac0-6ba9db04a50b" Dec 10 19:17:03 crc kubenswrapper[4828]: I1210 19:17:03.737763 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-dxcl8" event={"ID":"4246d7f6-cfb7-4d1d-aac0-6ba9db04a50b","Type":"ContainerStarted","Data":"060dfec1e972a7f491fd12fd2d08ee8535c51565157b8520996219881d056491"} Dec 10 19:17:03 crc kubenswrapper[4828]: I1210 19:17:03.738059 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-dxcl8" Dec 10 19:17:03 crc kubenswrapper[4828]: E1210 19:17:03.742316 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-dxcl8" podUID="4246d7f6-cfb7-4d1d-aac0-6ba9db04a50b" Dec 10 19:17:03 crc kubenswrapper[4828]: I1210 19:17:03.745491 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" event={"ID":"25aaba8d-1069-4360-a0b4-af30543dbb02","Type":"ContainerStarted","Data":"bb26b5c4b44efa06ac5a32f7e691ba36adf785b96cd166ad526b7ef94366db17"} Dec 10 19:17:03 crc kubenswrapper[4828]: I1210 19:17:03.746261 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:17:05 crc kubenswrapper[4828]: I1210 19:17:05.762315 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mmqkg" event={"ID":"bad66884-aa3d-4683-abc6-2387140d407f","Type":"ContainerStarted","Data":"db61487bd3d721cd2e522b7ccd2024b0e8423be5e30c9a27cbf69bce9e266862"} Dec 10 19:17:05 crc kubenswrapper[4828]: I1210 19:17:05.765365 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-wtcz4" event={"ID":"bd786401-638a-4830-8a28-cc01b03d311d","Type":"ContainerStarted","Data":"8b6bb8475aeb4b11cdd5c22e0689e1450aee0e8b04eb7bd71b7a950a74cc0ec7"} Dec 10 19:17:05 crc kubenswrapper[4828]: I1210 19:17:05.769348 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-njhhp" event={"ID":"8f635d23-992f-4a05-9896-c38d41ec853d","Type":"ContainerStarted","Data":"15e0b912561adc514253a906182eddf977c1f49799a459597a4138886f9f2a7c"} Dec 10 19:17:05 crc kubenswrapper[4828]: I1210 19:17:05.771013 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6bqjc" event={"ID":"23431602-592f-4c26-a5ab-1ada81fda48a","Type":"ContainerStarted","Data":"7b82c3ba912d75afa288c0a28394442c8d6845c257c8a7b3a9fa5965661de331"} Dec 10 19:17:06 crc kubenswrapper[4828]: I1210 19:17:06.805856 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" podStartSLOduration=76.80584079 podStartE2EDuration="1m16.80584079s" podCreationTimestamp="2025-12-10 19:15:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:17:03.802229394 +0000 UTC m=+1304.312840439" watchObservedRunningTime="2025-12-10 19:17:06.80584079 +0000 UTC m=+1307.316451795" Dec 10 19:17:07 crc kubenswrapper[4828]: E1210 19:17:07.341836 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-pcfhr" podUID="b0544042-b6b7-4a9d-884a-12859ddea597" Dec 10 19:17:07 crc kubenswrapper[4828]: E1210 19:17:07.742711 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg" podUID="e87fb7c5-676f-4549-82e3-2696306f651b" Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.818146 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-f7cn8" event={"ID":"85a6c448-c39f-4ec5-b1f7-4f7ff1bc43a0","Type":"ContainerStarted","Data":"4fb2be0e465840f308ab35e87dcbddafc218fcde903647428369e6311bb66c2b"} Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.826428 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-pcfhr" event={"ID":"b0544042-b6b7-4a9d-884a-12859ddea597","Type":"ContainerStarted","Data":"3070a3ef278dd9ec5227dc1cce49d03742f512870750f0cec4e2243a89c08cc8"} Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.841555 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pf44z" event={"ID":"65429c05-8bff-4545-a085-e4c7faa29fa3","Type":"ContainerStarted","Data":"3e3e57b8f5bfd6f54913aa73f59c3aac6f089b9f39ad61bfd265926f3b933767"} Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.842830 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pf44z" Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.849929 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-wtcz4" event={"ID":"bd786401-638a-4830-8a28-cc01b03d311d","Type":"ContainerStarted","Data":"6679675c2349ac7bdfbb4b8152f067b22f213fbbcca87a747420d93bcc37ea76"} Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.851181 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pf44z" Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.851529 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-wtcz4" Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.859345 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-njhhp" event={"ID":"8f635d23-992f-4a05-9896-c38d41ec853d","Type":"ContainerStarted","Data":"5047433228ba8029a04665616afcb7bdb4c6c2a61a6fdf2d9e932b49050e3b32"} Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.860311 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-njhhp" Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.865181 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6bqjc" event={"ID":"23431602-592f-4c26-a5ab-1ada81fda48a","Type":"ContainerStarted","Data":"9661c2cb309aec91bc575626828ddd7f21c61c75a4fc0c8a35ade495e405d0c3"} Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.865943 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6bqjc" Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.867160 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg" event={"ID":"e87fb7c5-676f-4549-82e3-2696306f651b","Type":"ContainerStarted","Data":"e005f730962f592c13ec0655dda0510d9bfc2e81b269c29cb9a98416d0d1e9f5"} Dec 10 19:17:07 crc kubenswrapper[4828]: E1210 19:17:07.891237 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg" podUID="e87fb7c5-676f-4549-82e3-2696306f651b" Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.894565 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mmqkg" event={"ID":"bad66884-aa3d-4683-abc6-2387140d407f","Type":"ContainerStarted","Data":"cc37e458d087968fddbe3807eac4a6f4523da45feed41f215b38de66b4efdda1"} Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.895311 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mmqkg" Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.903199 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-wtcz4" podStartSLOduration=7.456672715 podStartE2EDuration="1m18.903182301s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:15:51.724040064 +0000 UTC m=+1232.234651069" lastFinishedPulling="2025-12-10 19:17:03.17054965 +0000 UTC m=+1303.681160655" observedRunningTime="2025-12-10 19:17:07.891011672 +0000 UTC m=+1308.401622677" watchObservedRunningTime="2025-12-10 19:17:07.903182301 +0000 UTC m=+1308.413793306" Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.910076 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vhf9j" event={"ID":"1f3d156d-ddc6-47a2-9c40-5b675e535b37","Type":"ContainerStarted","Data":"752b99d54bc7c3ae22306683458dfc077eb6a0cad08cd6c7e87d616069350b76"} Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.964328 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pf44z" podStartSLOduration=4.202216479 podStartE2EDuration="1m18.964311913s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:15:52.43752184 +0000 UTC m=+1232.948132845" lastFinishedPulling="2025-12-10 19:17:07.199617274 +0000 UTC m=+1307.710228279" observedRunningTime="2025-12-10 19:17:07.92202977 +0000 UTC m=+1308.432640795" watchObservedRunningTime="2025-12-10 19:17:07.964311913 +0000 UTC m=+1308.474922918" Dec 10 19:17:07 crc kubenswrapper[4828]: I1210 19:17:07.968369 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mmqkg" podStartSLOduration=6.860717663 podStartE2EDuration="1m18.968359233s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:15:51.063174077 +0000 UTC m=+1231.573785072" lastFinishedPulling="2025-12-10 19:17:03.170815627 +0000 UTC m=+1303.681426642" observedRunningTime="2025-12-10 19:17:07.960407968 +0000 UTC m=+1308.471018973" watchObservedRunningTime="2025-12-10 19:17:07.968359233 +0000 UTC m=+1308.478970238" Dec 10 19:17:08 crc kubenswrapper[4828]: I1210 19:17:08.005823 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-njhhp" podStartSLOduration=7.930873517 podStartE2EDuration="1m19.005784244s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:15:51.680610766 +0000 UTC m=+1232.191221771" lastFinishedPulling="2025-12-10 19:17:02.755521493 +0000 UTC m=+1303.266132498" observedRunningTime="2025-12-10 19:17:07.985115585 +0000 UTC m=+1308.495726590" watchObservedRunningTime="2025-12-10 19:17:08.005784244 +0000 UTC m=+1308.516395249" Dec 10 19:17:08 crc kubenswrapper[4828]: I1210 19:17:08.035857 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6bqjc" podStartSLOduration=7.366808367 podStartE2EDuration="1m19.035836966s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:15:51.504238314 +0000 UTC m=+1232.014849309" lastFinishedPulling="2025-12-10 19:17:03.173266903 +0000 UTC m=+1303.683877908" observedRunningTime="2025-12-10 19:17:08.029833994 +0000 UTC m=+1308.540444999" watchObservedRunningTime="2025-12-10 19:17:08.035836966 +0000 UTC m=+1308.546447971" Dec 10 19:17:08 crc kubenswrapper[4828]: I1210 19:17:08.927596 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cv2bf" event={"ID":"0a273a13-6393-4a24-85c8-8ac01dbf0e5c","Type":"ContainerStarted","Data":"5d758bbf423ceae2c1ca8022a72b8f53e6a775d99a34de86a158fad0901f5c73"} Dec 10 19:17:08 crc kubenswrapper[4828]: I1210 19:17:08.960312 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cv2bf" podStartSLOduration=4.18630388 podStartE2EDuration="1m18.960283744s" podCreationTimestamp="2025-12-10 19:15:50 +0000 UTC" firstStartedPulling="2025-12-10 19:15:52.5581069 +0000 UTC m=+1233.068717905" lastFinishedPulling="2025-12-10 19:17:07.332086764 +0000 UTC m=+1307.842697769" observedRunningTime="2025-12-10 19:17:08.947867358 +0000 UTC m=+1309.458478363" watchObservedRunningTime="2025-12-10 19:17:08.960283744 +0000 UTC m=+1309.470894779" Dec 10 19:17:09 crc kubenswrapper[4828]: E1210 19:17:09.112416 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-88stb" podUID="3c945977-c07e-4ef4-9370-d2fa1264eee0" Dec 10 19:17:09 crc kubenswrapper[4828]: E1210 19:17:09.125965 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-cfhjs" podUID="8991af61-9866-4ae9-b7a7-7545833c4485" Dec 10 19:17:09 crc kubenswrapper[4828]: E1210 19:17:09.148319 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk" podUID="1c7fb44a-c32b-49f0-8220-9cb69be63935" Dec 10 19:17:09 crc kubenswrapper[4828]: E1210 19:17:09.158440 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-j64bx" podUID="70814aec-fe59-4dc6-a085-c5a5da7b93a5" Dec 10 19:17:09 crc kubenswrapper[4828]: E1210 19:17:09.172299 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ftlx" podUID="be987ffc-f75a-4364-ac29-37fbc9bb7200" Dec 10 19:17:09 crc kubenswrapper[4828]: E1210 19:17:09.393032 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-2z6fx" podUID="e0825966-a2fc-4a99-9fc8-d256f885a2c5" Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.937338 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" event={"ID":"55bb5bf6-cc44-43f9-84e5-1137401fd188","Type":"ContainerStarted","Data":"125460964c5aa592c40a6e6e59e6cdcc94d04e328cce021591c484652c0f0580"} Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.939776 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-79p7l" event={"ID":"be915c9b-8dfe-4aab-8e74-5375e365e623","Type":"ContainerStarted","Data":"441eed2a176ef352604ec8c285842b178ff83725457ef654c8eabeffef90dc6d"} Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.941128 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-79p7l" Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.941567 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk" event={"ID":"1c7fb44a-c32b-49f0-8220-9cb69be63935","Type":"ContainerStarted","Data":"c9333795cf793c914691bd239f3f455517e97eb4f4328dc800a72f1cb4fe6f3d"} Dec 10 19:17:09 crc kubenswrapper[4828]: E1210 19:17:09.943350 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:961417d59f527d925ac48ff6a11de747d0493315e496e34dc83d76a1a1fff58a\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk" podUID="1c7fb44a-c32b-49f0-8220-9cb69be63935" Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.943487 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-f7cn8" event={"ID":"85a6c448-c39f-4ec5-b1f7-4f7ff1bc43a0","Type":"ContainerStarted","Data":"724eab9c33ed6a432e743b8d7b872563cca5a31fb6d706e412db460bed606b8d"} Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.943537 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-f7cn8" Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.944705 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-79p7l" Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.945674 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-smtjq" event={"ID":"ab0dbc03-b62b-40c9-98b9-7bb192cddff3","Type":"ContainerStarted","Data":"8efdf15e3c270710a3344d1874ba58168b2afee6fcdaff7c99c572f7d89dfcb8"} Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.946294 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-smtjq" Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.948507 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-88stb" event={"ID":"3c945977-c07e-4ef4-9370-d2fa1264eee0","Type":"ContainerStarted","Data":"45db3ef10cb189d5be49ba9e7778ea7fcf9163d3fe5c4217a828f0ef705de7d1"} Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.948630 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-smtjq" Dec 10 19:17:09 crc kubenswrapper[4828]: E1210 19:17:09.950223 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\"" pod="openstack-operators/placement-operator-controller-manager-78f8948974-88stb" podUID="3c945977-c07e-4ef4-9370-d2fa1264eee0" Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.950299 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ftlx" event={"ID":"be987ffc-f75a-4364-ac29-37fbc9bb7200","Type":"ContainerStarted","Data":"6eb0de83c5e11c054d7abcd0637e613229e82be93c87cebaad8fc6649b5c59a2"} Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.952069 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-942ph" event={"ID":"d7200137-65c3-48b5-bc24-5cf17caa07f4","Type":"ContainerStarted","Data":"7f2872cdcd6b5981c5fbc49233059b96bd45d469ae3e8c014eb044ed857ca4b8"} Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.952364 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-942ph" Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.956969 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-j64bx" event={"ID":"70814aec-fe59-4dc6-a085-c5a5da7b93a5","Type":"ContainerStarted","Data":"d3625d392f6b2b363dd9c701d6ad6b35d73cda63d398376ab444277ca8acd880"} Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.957059 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-942ph" Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.961560 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vhf9j" event={"ID":"1f3d156d-ddc6-47a2-9c40-5b675e535b37","Type":"ContainerStarted","Data":"9026cc9dabc1e9d17656944f5d106668d46c072b1c376df2927679cc81820b87"} Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.962388 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vhf9j" Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.966549 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-79p7l" podStartSLOduration=4.845319091 podStartE2EDuration="1m20.966530962s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:15:51.492466605 +0000 UTC m=+1232.003077610" lastFinishedPulling="2025-12-10 19:17:07.613678466 +0000 UTC m=+1308.124289481" observedRunningTime="2025-12-10 19:17:09.962050211 +0000 UTC m=+1310.472661236" watchObservedRunningTime="2025-12-10 19:17:09.966530962 +0000 UTC m=+1310.477141977" Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.968316 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-cfhjs" event={"ID":"8991af61-9866-4ae9-b7a7-7545833c4485","Type":"ContainerStarted","Data":"bb120a9ed493eb68649e9ac335911517f90400bbd1f41734a7e109b970e70868"} Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.988018 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-dxcl8" event={"ID":"4246d7f6-cfb7-4d1d-aac0-6ba9db04a50b","Type":"ContainerStarted","Data":"cc92121878359a908d86a252cbc9b26dc8809767df13d732e281e3ad096301da"} Dec 10 19:17:09 crc kubenswrapper[4828]: I1210 19:17:09.994115 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-dxcl8" Dec 10 19:17:10 crc kubenswrapper[4828]: I1210 19:17:10.000546 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-pcfhr" event={"ID":"b0544042-b6b7-4a9d-884a-12859ddea597","Type":"ContainerStarted","Data":"a47407aecdeda6849e18c179ee27661cc044912de0624bff6cdf8b5efb0b9ab1"} Dec 10 19:17:10 crc kubenswrapper[4828]: I1210 19:17:10.000704 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-pcfhr" Dec 10 19:17:10 crc kubenswrapper[4828]: I1210 19:17:10.019001 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-2z6fx" event={"ID":"e0825966-a2fc-4a99-9fc8-d256f885a2c5","Type":"ContainerStarted","Data":"297d244b4692f8a9cd190116b42267c073f2edd17629f4e1d1acd72d9348df58"} Dec 10 19:17:10 crc kubenswrapper[4828]: I1210 19:17:10.029063 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" event={"ID":"4614b8b1-2082-4b05-8473-404f0f82772b","Type":"ContainerStarted","Data":"75da77d22c90a6c91c58c1a8c8c48772559530241f26e4f239c3f103610dc4c2"} Dec 10 19:17:10 crc kubenswrapper[4828]: I1210 19:17:10.032382 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-wtcz4" Dec 10 19:17:10 crc kubenswrapper[4828]: I1210 19:17:10.032611 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-njhhp" Dec 10 19:17:10 crc kubenswrapper[4828]: I1210 19:17:10.032662 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-6bqjc" Dec 10 19:17:10 crc kubenswrapper[4828]: I1210 19:17:10.034004 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-mmqkg" Dec 10 19:17:10 crc kubenswrapper[4828]: I1210 19:17:10.067576 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vhf9j" podStartSLOduration=8.962716824 podStartE2EDuration="1m21.067553082s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:15:51.066933775 +0000 UTC m=+1231.577544780" lastFinishedPulling="2025-12-10 19:17:03.171770043 +0000 UTC m=+1303.682381038" observedRunningTime="2025-12-10 19:17:10.012038012 +0000 UTC m=+1310.522649017" watchObservedRunningTime="2025-12-10 19:17:10.067553082 +0000 UTC m=+1310.578164087" Dec 10 19:17:10 crc kubenswrapper[4828]: I1210 19:17:10.076777 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-f7cn8" podStartSLOduration=9.384866004 podStartE2EDuration="1m21.076758751s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:15:51.503526645 +0000 UTC m=+1232.014137650" lastFinishedPulling="2025-12-10 19:17:03.195419392 +0000 UTC m=+1303.706030397" observedRunningTime="2025-12-10 19:17:10.060902203 +0000 UTC m=+1310.571513208" watchObservedRunningTime="2025-12-10 19:17:10.076758751 +0000 UTC m=+1310.587369766" Dec 10 19:17:10 crc kubenswrapper[4828]: I1210 19:17:10.093611 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-smtjq" podStartSLOduration=4.81683259 podStartE2EDuration="1m21.093592926s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:15:52.548755045 +0000 UTC m=+1233.059366050" lastFinishedPulling="2025-12-10 19:17:08.825515381 +0000 UTC m=+1309.336126386" observedRunningTime="2025-12-10 19:17:10.085108767 +0000 UTC m=+1310.595719782" watchObservedRunningTime="2025-12-10 19:17:10.093592926 +0000 UTC m=+1310.604203931" Dec 10 19:17:10 crc kubenswrapper[4828]: I1210 19:17:10.152730 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-942ph" podStartSLOduration=5.10831891 podStartE2EDuration="1m20.152708934s" podCreationTimestamp="2025-12-10 19:15:50 +0000 UTC" firstStartedPulling="2025-12-10 19:15:52.511015076 +0000 UTC m=+1233.021626081" lastFinishedPulling="2025-12-10 19:17:07.5554051 +0000 UTC m=+1308.066016105" observedRunningTime="2025-12-10 19:17:10.152344774 +0000 UTC m=+1310.662955779" watchObservedRunningTime="2025-12-10 19:17:10.152708934 +0000 UTC m=+1310.663319939" Dec 10 19:17:10 crc kubenswrapper[4828]: I1210 19:17:10.231703 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-dxcl8" podStartSLOduration=4.104397657 podStartE2EDuration="1m21.231675289s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:15:51.719868864 +0000 UTC m=+1232.230479869" lastFinishedPulling="2025-12-10 19:17:08.847146496 +0000 UTC m=+1309.357757501" observedRunningTime="2025-12-10 19:17:10.227315801 +0000 UTC m=+1310.737926806" watchObservedRunningTime="2025-12-10 19:17:10.231675289 +0000 UTC m=+1310.742286294" Dec 10 19:17:10 crc kubenswrapper[4828]: I1210 19:17:10.346257 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-pcfhr" podStartSLOduration=3.625756406 podStartE2EDuration="1m21.346236575s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:15:51.118445585 +0000 UTC m=+1231.629056580" lastFinishedPulling="2025-12-10 19:17:08.838925744 +0000 UTC m=+1309.349536749" observedRunningTime="2025-12-10 19:17:10.342263648 +0000 UTC m=+1310.852874653" watchObservedRunningTime="2025-12-10 19:17:10.346236575 +0000 UTC m=+1310.856847580" Dec 10 19:17:11 crc kubenswrapper[4828]: I1210 19:17:11.038750 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" event={"ID":"4614b8b1-2082-4b05-8473-404f0f82772b","Type":"ContainerStarted","Data":"9d3ac3aff9c88f3ae2ca3b92d1d6874be52e286733e6d98a41c1e9e2d1d8aa8b"} Dec 10 19:17:11 crc kubenswrapper[4828]: I1210 19:17:11.039922 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" Dec 10 19:17:11 crc kubenswrapper[4828]: I1210 19:17:11.041431 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-j64bx" event={"ID":"70814aec-fe59-4dc6-a085-c5a5da7b93a5","Type":"ContainerStarted","Data":"5de401701ca4a1d194647b39101c8c010790b5e33822815fad470d80a1c3f2ab"} Dec 10 19:17:11 crc kubenswrapper[4828]: I1210 19:17:11.041893 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-j64bx" Dec 10 19:17:11 crc kubenswrapper[4828]: I1210 19:17:11.048511 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-2z6fx" event={"ID":"e0825966-a2fc-4a99-9fc8-d256f885a2c5","Type":"ContainerStarted","Data":"e2a04969fb5c55ac4285f963338bac66354b03822ea5a573846010b20c995777"} Dec 10 19:17:11 crc kubenswrapper[4828]: I1210 19:17:11.048618 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-2z6fx" Dec 10 19:17:11 crc kubenswrapper[4828]: I1210 19:17:11.052612 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" event={"ID":"55bb5bf6-cc44-43f9-84e5-1137401fd188","Type":"ContainerStarted","Data":"a5fa5ac24d86524817d7f4ec29a66eb85542c9dea3a55de83b56fda44978c82f"} Dec 10 19:17:11 crc kubenswrapper[4828]: I1210 19:17:11.052845 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" Dec 10 19:17:11 crc kubenswrapper[4828]: I1210 19:17:11.063188 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" podStartSLOduration=74.317769409 podStartE2EDuration="1m22.063171204s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:17:00.962923649 +0000 UTC m=+1301.473534654" lastFinishedPulling="2025-12-10 19:17:08.708325444 +0000 UTC m=+1309.218936449" observedRunningTime="2025-12-10 19:17:11.060765299 +0000 UTC m=+1311.571376304" watchObservedRunningTime="2025-12-10 19:17:11.063171204 +0000 UTC m=+1311.573782209" Dec 10 19:17:11 crc kubenswrapper[4828]: I1210 19:17:11.084460 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-j64bx" podStartSLOduration=3.928562622 podStartE2EDuration="1m22.084440279s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:15:52.54854056 +0000 UTC m=+1233.059151565" lastFinishedPulling="2025-12-10 19:17:10.704418217 +0000 UTC m=+1311.215029222" observedRunningTime="2025-12-10 19:17:11.081639403 +0000 UTC m=+1311.592250428" watchObservedRunningTime="2025-12-10 19:17:11.084440279 +0000 UTC m=+1311.595051294" Dec 10 19:17:11 crc kubenswrapper[4828]: I1210 19:17:11.149653 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" podStartSLOduration=74.261543699 podStartE2EDuration="1m22.149631111s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:17:00.959050634 +0000 UTC m=+1301.469661639" lastFinishedPulling="2025-12-10 19:17:08.847138046 +0000 UTC m=+1309.357749051" observedRunningTime="2025-12-10 19:17:11.147295198 +0000 UTC m=+1311.657906203" watchObservedRunningTime="2025-12-10 19:17:11.149631111 +0000 UTC m=+1311.660242126" Dec 10 19:17:11 crc kubenswrapper[4828]: I1210 19:17:11.208669 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-2z6fx" podStartSLOduration=3.384038473 podStartE2EDuration="1m21.208648266s" podCreationTimestamp="2025-12-10 19:15:50 +0000 UTC" firstStartedPulling="2025-12-10 19:15:52.547764139 +0000 UTC m=+1233.058375144" lastFinishedPulling="2025-12-10 19:17:10.372373932 +0000 UTC m=+1310.882984937" observedRunningTime="2025-12-10 19:17:11.200954348 +0000 UTC m=+1311.711565373" watchObservedRunningTime="2025-12-10 19:17:11.208648266 +0000 UTC m=+1311.719259271" Dec 10 19:17:12 crc kubenswrapper[4828]: I1210 19:17:12.061032 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-cfhjs" event={"ID":"8991af61-9866-4ae9-b7a7-7545833c4485","Type":"ContainerStarted","Data":"e63c15170a138fa9955611b461dfc3aa8319ee2afc2bfa62383dde5ad9f9ed74"} Dec 10 19:17:12 crc kubenswrapper[4828]: I1210 19:17:12.061421 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-cfhjs" Dec 10 19:17:12 crc kubenswrapper[4828]: I1210 19:17:12.063303 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ftlx" event={"ID":"be987ffc-f75a-4364-ac29-37fbc9bb7200","Type":"ContainerStarted","Data":"4262fa89f93336cb8a02f43c52e7187bbf9c6349412a4e9cae598a29fbcfb27f"} Dec 10 19:17:12 crc kubenswrapper[4828]: I1210 19:17:12.065235 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-vhf9j" Dec 10 19:17:12 crc kubenswrapper[4828]: I1210 19:17:12.081410 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-cfhjs" podStartSLOduration=3.9617642440000003 podStartE2EDuration="1m23.081391816s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:15:51.693349839 +0000 UTC m=+1232.203960844" lastFinishedPulling="2025-12-10 19:17:10.812977421 +0000 UTC m=+1311.323588416" observedRunningTime="2025-12-10 19:17:12.077565913 +0000 UTC m=+1312.588176918" watchObservedRunningTime="2025-12-10 19:17:12.081391816 +0000 UTC m=+1312.592002811" Dec 10 19:17:12 crc kubenswrapper[4828]: I1210 19:17:12.112556 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ftlx" podStartSLOduration=4.770788695 podStartE2EDuration="1m23.112540038s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:15:52.501733323 +0000 UTC m=+1233.012344328" lastFinishedPulling="2025-12-10 19:17:10.843484666 +0000 UTC m=+1311.354095671" observedRunningTime="2025-12-10 19:17:12.111414738 +0000 UTC m=+1312.622025743" watchObservedRunningTime="2025-12-10 19:17:12.112540038 +0000 UTC m=+1312.623151043" Dec 10 19:17:13 crc kubenswrapper[4828]: I1210 19:17:13.071288 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ftlx" Dec 10 19:17:16 crc kubenswrapper[4828]: I1210 19:17:16.023850 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-x27kt" Dec 10 19:17:16 crc kubenswrapper[4828]: I1210 19:17:16.039151 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fk7hbh" Dec 10 19:17:17 crc kubenswrapper[4828]: I1210 19:17:17.130822 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7b5585cf46-6j82f" Dec 10 19:17:18 crc kubenswrapper[4828]: I1210 19:17:18.792364 4828 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:17:19 crc kubenswrapper[4828]: I1210 19:17:19.967750 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-pcfhr" Dec 10 19:17:20 crc kubenswrapper[4828]: I1210 19:17:20.131211 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg" event={"ID":"e87fb7c5-676f-4549-82e3-2696306f651b","Type":"ContainerStarted","Data":"b3dadeb1c7676bf750b0119fb2e045c9459c1ec7d212b12dfa9b759e45f253f8"} Dec 10 19:17:20 crc kubenswrapper[4828]: I1210 19:17:20.131692 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg" Dec 10 19:17:20 crc kubenswrapper[4828]: I1210 19:17:20.146306 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg" podStartSLOduration=4.034173321 podStartE2EDuration="1m31.146287107s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:15:52.615584247 +0000 UTC m=+1233.126195252" lastFinishedPulling="2025-12-10 19:17:19.727698033 +0000 UTC m=+1320.238309038" observedRunningTime="2025-12-10 19:17:20.144851328 +0000 UTC m=+1320.655462333" watchObservedRunningTime="2025-12-10 19:17:20.146287107 +0000 UTC m=+1320.656898112" Dec 10 19:17:20 crc kubenswrapper[4828]: I1210 19:17:20.271786 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-f7cn8" Dec 10 19:17:20 crc kubenswrapper[4828]: I1210 19:17:20.287599 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-cfhjs" Dec 10 19:17:20 crc kubenswrapper[4828]: I1210 19:17:20.532884 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-4ftlx" Dec 10 19:17:20 crc kubenswrapper[4828]: I1210 19:17:20.660467 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-j64bx" Dec 10 19:17:21 crc kubenswrapper[4828]: I1210 19:17:21.016679 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-954d6f856-2z6fx" Dec 10 19:17:21 crc kubenswrapper[4828]: I1210 19:17:21.230851 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:17:21 crc kubenswrapper[4828]: I1210 19:17:21.230922 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:17:24 crc kubenswrapper[4828]: I1210 19:17:24.276651 4828 scope.go:117] "RemoveContainer" containerID="e1de59dac8272736eefcba4bf88453d4a852b5d12c59810902a1373b38571b69" Dec 10 19:17:24 crc kubenswrapper[4828]: I1210 19:17:24.299136 4828 scope.go:117] "RemoveContainer" containerID="e17c44148dab8622b13de1c71a7e901317109e245eaecad0fd3b4301c53c28c1" Dec 10 19:17:24 crc kubenswrapper[4828]: I1210 19:17:24.328135 4828 scope.go:117] "RemoveContainer" containerID="4d409d3cf1a68c6992d396dae4551433001fea8594dbf317e4f476da2b58b141" Dec 10 19:17:26 crc kubenswrapper[4828]: I1210 19:17:26.179947 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk" event={"ID":"1c7fb44a-c32b-49f0-8220-9cb69be63935","Type":"ContainerStarted","Data":"c9c1af5f8f1a8fe8d992f99fdd5f19a423566489bcce91cef0cd5a77435f71e7"} Dec 10 19:17:26 crc kubenswrapper[4828]: I1210 19:17:26.180588 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk" Dec 10 19:17:26 crc kubenswrapper[4828]: I1210 19:17:26.181678 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-88stb" event={"ID":"3c945977-c07e-4ef4-9370-d2fa1264eee0","Type":"ContainerStarted","Data":"600072a8d7a05a46752ffc1e2dc78f2eaaa66cc61e417700a7409a522e0fe2e3"} Dec 10 19:17:26 crc kubenswrapper[4828]: I1210 19:17:26.181971 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-88stb" Dec 10 19:17:26 crc kubenswrapper[4828]: I1210 19:17:26.208692 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk" podStartSLOduration=3.123710595 podStartE2EDuration="1m36.208675079s" podCreationTimestamp="2025-12-10 19:15:50 +0000 UTC" firstStartedPulling="2025-12-10 19:15:52.634965114 +0000 UTC m=+1233.145576119" lastFinishedPulling="2025-12-10 19:17:25.719929588 +0000 UTC m=+1326.230540603" observedRunningTime="2025-12-10 19:17:26.200389435 +0000 UTC m=+1326.711000430" watchObservedRunningTime="2025-12-10 19:17:26.208675079 +0000 UTC m=+1326.719286084" Dec 10 19:17:26 crc kubenswrapper[4828]: I1210 19:17:26.224078 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-88stb" podStartSLOduration=4.06087193 podStartE2EDuration="1m37.224058375s" podCreationTimestamp="2025-12-10 19:15:49 +0000 UTC" firstStartedPulling="2025-12-10 19:15:52.558424389 +0000 UTC m=+1233.069035394" lastFinishedPulling="2025-12-10 19:17:25.721610834 +0000 UTC m=+1326.232221839" observedRunningTime="2025-12-10 19:17:26.214676161 +0000 UTC m=+1326.725287166" watchObservedRunningTime="2025-12-10 19:17:26.224058375 +0000 UTC m=+1326.734669380" Dec 10 19:17:30 crc kubenswrapper[4828]: I1210 19:17:30.515911 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-s5jjg" Dec 10 19:17:30 crc kubenswrapper[4828]: I1210 19:17:30.751542 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-88stb" Dec 10 19:17:31 crc kubenswrapper[4828]: I1210 19:17:31.179509 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-75944c9b7-s84vk" Dec 10 19:17:47 crc kubenswrapper[4828]: I1210 19:17:47.843871 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-znhwl"] Dec 10 19:17:47 crc kubenswrapper[4828]: I1210 19:17:47.845905 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-znhwl" Dec 10 19:17:47 crc kubenswrapper[4828]: I1210 19:17:47.851675 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 10 19:17:47 crc kubenswrapper[4828]: I1210 19:17:47.851822 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-cz8gc" Dec 10 19:17:47 crc kubenswrapper[4828]: I1210 19:17:47.851962 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 10 19:17:47 crc kubenswrapper[4828]: I1210 19:17:47.852580 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 10 19:17:47 crc kubenswrapper[4828]: I1210 19:17:47.863148 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-znhwl"] Dec 10 19:17:47 crc kubenswrapper[4828]: I1210 19:17:47.917696 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2wdrq"] Dec 10 19:17:47 crc kubenswrapper[4828]: I1210 19:17:47.919240 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2wdrq" Dec 10 19:17:47 crc kubenswrapper[4828]: I1210 19:17:47.922718 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 10 19:17:47 crc kubenswrapper[4828]: I1210 19:17:47.924130 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21b70614-6116-47aa-8336-afb75469d5ac-config\") pod \"dnsmasq-dns-675f4bcbfc-znhwl\" (UID: \"21b70614-6116-47aa-8336-afb75469d5ac\") " pod="openstack/dnsmasq-dns-675f4bcbfc-znhwl" Dec 10 19:17:47 crc kubenswrapper[4828]: I1210 19:17:47.924222 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rgjt\" (UniqueName: \"kubernetes.io/projected/21b70614-6116-47aa-8336-afb75469d5ac-kube-api-access-6rgjt\") pod \"dnsmasq-dns-675f4bcbfc-znhwl\" (UID: \"21b70614-6116-47aa-8336-afb75469d5ac\") " pod="openstack/dnsmasq-dns-675f4bcbfc-znhwl" Dec 10 19:17:47 crc kubenswrapper[4828]: I1210 19:17:47.930795 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2wdrq"] Dec 10 19:17:48 crc kubenswrapper[4828]: I1210 19:17:48.025739 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21b70614-6116-47aa-8336-afb75469d5ac-config\") pod \"dnsmasq-dns-675f4bcbfc-znhwl\" (UID: \"21b70614-6116-47aa-8336-afb75469d5ac\") " pod="openstack/dnsmasq-dns-675f4bcbfc-znhwl" Dec 10 19:17:48 crc kubenswrapper[4828]: I1210 19:17:48.025840 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrmhx\" (UniqueName: \"kubernetes.io/projected/cb237099-ff45-4060-bb7d-d506f069837f-kube-api-access-xrmhx\") pod \"dnsmasq-dns-78dd6ddcc-2wdrq\" (UID: \"cb237099-ff45-4060-bb7d-d506f069837f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2wdrq" Dec 10 19:17:48 crc kubenswrapper[4828]: I1210 19:17:48.025884 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rgjt\" (UniqueName: \"kubernetes.io/projected/21b70614-6116-47aa-8336-afb75469d5ac-kube-api-access-6rgjt\") pod \"dnsmasq-dns-675f4bcbfc-znhwl\" (UID: \"21b70614-6116-47aa-8336-afb75469d5ac\") " pod="openstack/dnsmasq-dns-675f4bcbfc-znhwl" Dec 10 19:17:48 crc kubenswrapper[4828]: I1210 19:17:48.025919 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb237099-ff45-4060-bb7d-d506f069837f-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2wdrq\" (UID: \"cb237099-ff45-4060-bb7d-d506f069837f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2wdrq" Dec 10 19:17:48 crc kubenswrapper[4828]: I1210 19:17:48.026066 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb237099-ff45-4060-bb7d-d506f069837f-config\") pod \"dnsmasq-dns-78dd6ddcc-2wdrq\" (UID: \"cb237099-ff45-4060-bb7d-d506f069837f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2wdrq" Dec 10 19:17:48 crc kubenswrapper[4828]: I1210 19:17:48.026668 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21b70614-6116-47aa-8336-afb75469d5ac-config\") pod \"dnsmasq-dns-675f4bcbfc-znhwl\" (UID: \"21b70614-6116-47aa-8336-afb75469d5ac\") " pod="openstack/dnsmasq-dns-675f4bcbfc-znhwl" Dec 10 19:17:48 crc kubenswrapper[4828]: I1210 19:17:48.055736 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rgjt\" (UniqueName: \"kubernetes.io/projected/21b70614-6116-47aa-8336-afb75469d5ac-kube-api-access-6rgjt\") pod \"dnsmasq-dns-675f4bcbfc-znhwl\" (UID: \"21b70614-6116-47aa-8336-afb75469d5ac\") " pod="openstack/dnsmasq-dns-675f4bcbfc-znhwl" Dec 10 19:17:48 crc kubenswrapper[4828]: I1210 19:17:48.127756 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb237099-ff45-4060-bb7d-d506f069837f-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2wdrq\" (UID: \"cb237099-ff45-4060-bb7d-d506f069837f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2wdrq" Dec 10 19:17:48 crc kubenswrapper[4828]: I1210 19:17:48.127859 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb237099-ff45-4060-bb7d-d506f069837f-config\") pod \"dnsmasq-dns-78dd6ddcc-2wdrq\" (UID: \"cb237099-ff45-4060-bb7d-d506f069837f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2wdrq" Dec 10 19:17:48 crc kubenswrapper[4828]: I1210 19:17:48.128012 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrmhx\" (UniqueName: \"kubernetes.io/projected/cb237099-ff45-4060-bb7d-d506f069837f-kube-api-access-xrmhx\") pod \"dnsmasq-dns-78dd6ddcc-2wdrq\" (UID: \"cb237099-ff45-4060-bb7d-d506f069837f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2wdrq" Dec 10 19:17:48 crc kubenswrapper[4828]: I1210 19:17:48.128838 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb237099-ff45-4060-bb7d-d506f069837f-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-2wdrq\" (UID: \"cb237099-ff45-4060-bb7d-d506f069837f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2wdrq" Dec 10 19:17:48 crc kubenswrapper[4828]: I1210 19:17:48.128880 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb237099-ff45-4060-bb7d-d506f069837f-config\") pod \"dnsmasq-dns-78dd6ddcc-2wdrq\" (UID: \"cb237099-ff45-4060-bb7d-d506f069837f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2wdrq" Dec 10 19:17:48 crc kubenswrapper[4828]: I1210 19:17:48.169170 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrmhx\" (UniqueName: \"kubernetes.io/projected/cb237099-ff45-4060-bb7d-d506f069837f-kube-api-access-xrmhx\") pod \"dnsmasq-dns-78dd6ddcc-2wdrq\" (UID: \"cb237099-ff45-4060-bb7d-d506f069837f\") " pod="openstack/dnsmasq-dns-78dd6ddcc-2wdrq" Dec 10 19:17:48 crc kubenswrapper[4828]: I1210 19:17:48.170522 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-znhwl" Dec 10 19:17:48 crc kubenswrapper[4828]: I1210 19:17:48.239859 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2wdrq" Dec 10 19:17:48 crc kubenswrapper[4828]: I1210 19:17:48.657539 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-znhwl"] Dec 10 19:17:48 crc kubenswrapper[4828]: I1210 19:17:48.866744 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2wdrq"] Dec 10 19:17:49 crc kubenswrapper[4828]: I1210 19:17:49.369277 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-znhwl" event={"ID":"21b70614-6116-47aa-8336-afb75469d5ac","Type":"ContainerStarted","Data":"011300d0f201745e8d832733121e9bbd3c99815c3a89705b93af0e226ac0a6ad"} Dec 10 19:17:49 crc kubenswrapper[4828]: I1210 19:17:49.370395 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-2wdrq" event={"ID":"cb237099-ff45-4060-bb7d-d506f069837f","Type":"ContainerStarted","Data":"bdee8b24f9ae37b0a37ee11ff65e6fbda928f19182c9b09263fa51b58779ece0"} Dec 10 19:17:50 crc kubenswrapper[4828]: I1210 19:17:50.790139 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-znhwl"] Dec 10 19:17:50 crc kubenswrapper[4828]: I1210 19:17:50.829684 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-xkgl6"] Dec 10 19:17:50 crc kubenswrapper[4828]: I1210 19:17:50.831652 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-xkgl6" Dec 10 19:17:50 crc kubenswrapper[4828]: I1210 19:17:50.840062 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-xkgl6"] Dec 10 19:17:50 crc kubenswrapper[4828]: I1210 19:17:50.872340 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebe57772-c202-4d69-986f-aa63f6eeed94-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-xkgl6\" (UID: \"ebe57772-c202-4d69-986f-aa63f6eeed94\") " pod="openstack/dnsmasq-dns-5ccc8479f9-xkgl6" Dec 10 19:17:50 crc kubenswrapper[4828]: I1210 19:17:50.872395 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebe57772-c202-4d69-986f-aa63f6eeed94-config\") pod \"dnsmasq-dns-5ccc8479f9-xkgl6\" (UID: \"ebe57772-c202-4d69-986f-aa63f6eeed94\") " pod="openstack/dnsmasq-dns-5ccc8479f9-xkgl6" Dec 10 19:17:50 crc kubenswrapper[4828]: I1210 19:17:50.872423 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k6gf\" (UniqueName: \"kubernetes.io/projected/ebe57772-c202-4d69-986f-aa63f6eeed94-kube-api-access-6k6gf\") pod \"dnsmasq-dns-5ccc8479f9-xkgl6\" (UID: \"ebe57772-c202-4d69-986f-aa63f6eeed94\") " pod="openstack/dnsmasq-dns-5ccc8479f9-xkgl6" Dec 10 19:17:50 crc kubenswrapper[4828]: I1210 19:17:50.974479 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebe57772-c202-4d69-986f-aa63f6eeed94-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-xkgl6\" (UID: \"ebe57772-c202-4d69-986f-aa63f6eeed94\") " pod="openstack/dnsmasq-dns-5ccc8479f9-xkgl6" Dec 10 19:17:50 crc kubenswrapper[4828]: I1210 19:17:50.974867 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebe57772-c202-4d69-986f-aa63f6eeed94-config\") pod \"dnsmasq-dns-5ccc8479f9-xkgl6\" (UID: \"ebe57772-c202-4d69-986f-aa63f6eeed94\") " pod="openstack/dnsmasq-dns-5ccc8479f9-xkgl6" Dec 10 19:17:50 crc kubenswrapper[4828]: I1210 19:17:50.974903 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k6gf\" (UniqueName: \"kubernetes.io/projected/ebe57772-c202-4d69-986f-aa63f6eeed94-kube-api-access-6k6gf\") pod \"dnsmasq-dns-5ccc8479f9-xkgl6\" (UID: \"ebe57772-c202-4d69-986f-aa63f6eeed94\") " pod="openstack/dnsmasq-dns-5ccc8479f9-xkgl6" Dec 10 19:17:50 crc kubenswrapper[4828]: I1210 19:17:50.976127 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebe57772-c202-4d69-986f-aa63f6eeed94-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-xkgl6\" (UID: \"ebe57772-c202-4d69-986f-aa63f6eeed94\") " pod="openstack/dnsmasq-dns-5ccc8479f9-xkgl6" Dec 10 19:17:50 crc kubenswrapper[4828]: I1210 19:17:50.976697 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebe57772-c202-4d69-986f-aa63f6eeed94-config\") pod \"dnsmasq-dns-5ccc8479f9-xkgl6\" (UID: \"ebe57772-c202-4d69-986f-aa63f6eeed94\") " pod="openstack/dnsmasq-dns-5ccc8479f9-xkgl6" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:50.998135 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k6gf\" (UniqueName: \"kubernetes.io/projected/ebe57772-c202-4d69-986f-aa63f6eeed94-kube-api-access-6k6gf\") pod \"dnsmasq-dns-5ccc8479f9-xkgl6\" (UID: \"ebe57772-c202-4d69-986f-aa63f6eeed94\") " pod="openstack/dnsmasq-dns-5ccc8479f9-xkgl6" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.081368 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2wdrq"] Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.110329 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4cpx8"] Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.113810 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.142574 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4cpx8"] Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.164746 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-xkgl6" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.234975 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.235038 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.235084 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.237748 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"04946921ee0aa69a9d56354bf922c69f57770e0a2b673248523db9ef4278fd20"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.237851 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://04946921ee0aa69a9d56354bf922c69f57770e0a2b673248523db9ef4278fd20" gracePeriod=600 Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.292650 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-4cpx8\" (UID: \"d2630e12-86b3-45d3-84c0-dd50f5cf8c15\") " pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.292782 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sl49\" (UniqueName: \"kubernetes.io/projected/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-kube-api-access-5sl49\") pod \"dnsmasq-dns-57d769cc4f-4cpx8\" (UID: \"d2630e12-86b3-45d3-84c0-dd50f5cf8c15\") " pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.292848 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-config\") pod \"dnsmasq-dns-57d769cc4f-4cpx8\" (UID: \"d2630e12-86b3-45d3-84c0-dd50f5cf8c15\") " pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.399829 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-4cpx8\" (UID: \"d2630e12-86b3-45d3-84c0-dd50f5cf8c15\") " pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.400299 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sl49\" (UniqueName: \"kubernetes.io/projected/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-kube-api-access-5sl49\") pod \"dnsmasq-dns-57d769cc4f-4cpx8\" (UID: \"d2630e12-86b3-45d3-84c0-dd50f5cf8c15\") " pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.400369 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-config\") pod \"dnsmasq-dns-57d769cc4f-4cpx8\" (UID: \"d2630e12-86b3-45d3-84c0-dd50f5cf8c15\") " pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.402109 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-4cpx8\" (UID: \"d2630e12-86b3-45d3-84c0-dd50f5cf8c15\") " pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.402992 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-config\") pod \"dnsmasq-dns-57d769cc4f-4cpx8\" (UID: \"d2630e12-86b3-45d3-84c0-dd50f5cf8c15\") " pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.410618 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="04946921ee0aa69a9d56354bf922c69f57770e0a2b673248523db9ef4278fd20" exitCode=0 Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.410655 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"04946921ee0aa69a9d56354bf922c69f57770e0a2b673248523db9ef4278fd20"} Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.410686 4828 scope.go:117] "RemoveContainer" containerID="6e3402faca982d603d83c3dbbde8307326fa3a18260295ae4bfb613477d32c1e" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.435040 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sl49\" (UniqueName: \"kubernetes.io/projected/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-kube-api-access-5sl49\") pod \"dnsmasq-dns-57d769cc4f-4cpx8\" (UID: \"d2630e12-86b3-45d3-84c0-dd50f5cf8c15\") " pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.456408 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.723956 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-xkgl6"] Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.938476 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.940189 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.943245 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.943618 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.943737 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.943961 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-zttjq" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.944074 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.944186 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.944336 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 10 19:17:51 crc kubenswrapper[4828]: I1210 19:17:51.953181 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.098934 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4cpx8"] Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.131057 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.131109 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.131145 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.131167 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fd6f7666-6cc8-478d-9acf-21cfbad541c9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.131190 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.131235 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.131254 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.131267 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.131281 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hftq\" (UniqueName: \"kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-kube-api-access-7hftq\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.131339 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.131378 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fd6f7666-6cc8-478d-9acf-21cfbad541c9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.235291 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.235341 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fd6f7666-6cc8-478d-9acf-21cfbad541c9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.235369 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.235424 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.235445 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.235461 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.235481 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hftq\" (UniqueName: \"kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-kube-api-access-7hftq\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.235531 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.235575 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fd6f7666-6cc8-478d-9acf-21cfbad541c9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.235594 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.235617 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.236511 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.236651 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.236728 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.237064 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.237091 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.243371 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fd6f7666-6cc8-478d-9acf-21cfbad541c9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.251136 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.252121 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fd6f7666-6cc8-478d-9acf-21cfbad541c9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.252198 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.252725 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.256279 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hftq\" (UniqueName: \"kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-kube-api-access-7hftq\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.258786 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.260939 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.268375 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-96rhn" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.269436 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.269722 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.269904 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.270047 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.270184 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.270211 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.287046 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.304969 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.426438 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-xkgl6" event={"ID":"ebe57772-c202-4d69-986f-aa63f6eeed94","Type":"ContainerStarted","Data":"aa2d2d10ea00fb51152a67173715a9da8f05237af7d980edcd00235ea39da53a"} Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.430128 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca"} Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.431585 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" event={"ID":"d2630e12-86b3-45d3-84c0-dd50f5cf8c15","Type":"ContainerStarted","Data":"51e476d61488997d2c755a2a84bffd103a0f79f37735a8f8aa9875553841b706"} Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.439700 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.439740 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b24f93de-d604-499b-be8e-2e934d119192-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.439759 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b24f93de-d604-499b-be8e-2e934d119192-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.439836 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.439853 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.439870 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.439889 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.439907 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-config-data\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.439929 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jxn8\" (UniqueName: \"kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-kube-api-access-6jxn8\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.439974 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.440006 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.541768 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.541829 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b24f93de-d604-499b-be8e-2e934d119192-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.541854 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b24f93de-d604-499b-be8e-2e934d119192-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.541904 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.541924 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.541951 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.541979 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.541992 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-config-data\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.542013 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jxn8\" (UniqueName: \"kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-kube-api-access-6jxn8\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.542063 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.542087 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.542968 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.543079 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.543698 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.544028 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.545154 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-config-data\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.545220 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.547481 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b24f93de-d604-499b-be8e-2e934d119192-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.548056 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.550238 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.554602 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b24f93de-d604-499b-be8e-2e934d119192-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.566517 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.567180 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jxn8\" (UniqueName: \"kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-kube-api-access-6jxn8\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.586248 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " pod="openstack/rabbitmq-server-0" Dec 10 19:17:52 crc kubenswrapper[4828]: I1210 19:17:52.654562 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.099740 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.283023 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:17:53 crc kubenswrapper[4828]: W1210 19:17:53.310849 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb24f93de_d604_499b_be8e_2e934d119192.slice/crio-81896e0a70a4edf55b884889ee4445be770f759065a5aa263cbeb0bda75df76c WatchSource:0}: Error finding container 81896e0a70a4edf55b884889ee4445be770f759065a5aa263cbeb0bda75df76c: Status 404 returned error can't find the container with id 81896e0a70a4edf55b884889ee4445be770f759065a5aa263cbeb0bda75df76c Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.449893 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fd6f7666-6cc8-478d-9acf-21cfbad541c9","Type":"ContainerStarted","Data":"4680091c3e0313bdbe5d56a5292605397c48c1f482b86a0c2b6c8c0514d3d36d"} Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.453732 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b24f93de-d604-499b-be8e-2e934d119192","Type":"ContainerStarted","Data":"81896e0a70a4edf55b884889ee4445be770f759065a5aa263cbeb0bda75df76c"} Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.894948 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.897720 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.900707 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.900994 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-2nxff" Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.901178 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.901218 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.912312 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.915591 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.978072 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ddf0d133-f122-4bac-b746-222a86e29dd2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.978153 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf0d133-f122-4bac-b746-222a86e29dd2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.978210 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ddf0d133-f122-4bac-b746-222a86e29dd2-kolla-config\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.978241 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddf0d133-f122-4bac-b746-222a86e29dd2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.978379 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ddf0d133-f122-4bac-b746-222a86e29dd2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.978406 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv9vf\" (UniqueName: \"kubernetes.io/projected/ddf0d133-f122-4bac-b746-222a86e29dd2-kube-api-access-xv9vf\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.978481 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:53 crc kubenswrapper[4828]: I1210 19:17:53.978540 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ddf0d133-f122-4bac-b746-222a86e29dd2-config-data-default\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.081569 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ddf0d133-f122-4bac-b746-222a86e29dd2-kolla-config\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.081639 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddf0d133-f122-4bac-b746-222a86e29dd2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.081690 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ddf0d133-f122-4bac-b746-222a86e29dd2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.081721 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv9vf\" (UniqueName: \"kubernetes.io/projected/ddf0d133-f122-4bac-b746-222a86e29dd2-kube-api-access-xv9vf\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.081771 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.081835 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ddf0d133-f122-4bac-b746-222a86e29dd2-config-data-default\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.081928 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ddf0d133-f122-4bac-b746-222a86e29dd2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.081981 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf0d133-f122-4bac-b746-222a86e29dd2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.082733 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.082929 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ddf0d133-f122-4bac-b746-222a86e29dd2-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.083576 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ddf0d133-f122-4bac-b746-222a86e29dd2-kolla-config\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.085032 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ddf0d133-f122-4bac-b746-222a86e29dd2-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.093825 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ddf0d133-f122-4bac-b746-222a86e29dd2-config-data-default\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.098494 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddf0d133-f122-4bac-b746-222a86e29dd2-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.116713 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf0d133-f122-4bac-b746-222a86e29dd2-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.118206 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.122088 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv9vf\" (UniqueName: \"kubernetes.io/projected/ddf0d133-f122-4bac-b746-222a86e29dd2-kube-api-access-xv9vf\") pod \"openstack-galera-0\" (UID: \"ddf0d133-f122-4bac-b746-222a86e29dd2\") " pod="openstack/openstack-galera-0" Dec 10 19:17:54 crc kubenswrapper[4828]: I1210 19:17:54.221074 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.324518 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.327846 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.337831 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-7lknc" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.338969 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.339008 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.339629 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.359390 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.419536 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8b0f0962-54b6-46c3-8c5d-d008874f754f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.419612 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b0f0962-54b6-46c3-8c5d-d008874f754f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.419639 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b0f0962-54b6-46c3-8c5d-d008874f754f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.419662 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkzdh\" (UniqueName: \"kubernetes.io/projected/8b0f0962-54b6-46c3-8c5d-d008874f754f-kube-api-access-wkzdh\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.419697 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8b0f0962-54b6-46c3-8c5d-d008874f754f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.419726 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b0f0962-54b6-46c3-8c5d-d008874f754f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.419742 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8b0f0962-54b6-46c3-8c5d-d008874f754f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.419784 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.521835 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8b0f0962-54b6-46c3-8c5d-d008874f754f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.521905 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b0f0962-54b6-46c3-8c5d-d008874f754f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.521936 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b0f0962-54b6-46c3-8c5d-d008874f754f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.521963 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkzdh\" (UniqueName: \"kubernetes.io/projected/8b0f0962-54b6-46c3-8c5d-d008874f754f-kube-api-access-wkzdh\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.521986 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8b0f0962-54b6-46c3-8c5d-d008874f754f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.522011 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b0f0962-54b6-46c3-8c5d-d008874f754f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.522027 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8b0f0962-54b6-46c3-8c5d-d008874f754f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.522051 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.522419 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.523677 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8b0f0962-54b6-46c3-8c5d-d008874f754f-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.523683 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b0f0962-54b6-46c3-8c5d-d008874f754f-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.523789 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8b0f0962-54b6-46c3-8c5d-d008874f754f-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.524531 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8b0f0962-54b6-46c3-8c5d-d008874f754f-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.530833 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b0f0962-54b6-46c3-8c5d-d008874f754f-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.538970 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b0f0962-54b6-46c3-8c5d-d008874f754f-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.542089 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkzdh\" (UniqueName: \"kubernetes.io/projected/8b0f0962-54b6-46c3-8c5d-d008874f754f-kube-api-access-wkzdh\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.554124 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"8b0f0962-54b6-46c3-8c5d-d008874f754f\") " pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.657449 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.667534 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.670562 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.676436 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-klsl2" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.676721 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.677040 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.692857 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.726395 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec-kolla-config\") pod \"memcached-0\" (UID: \"0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec\") " pod="openstack/memcached-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.726746 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znxnb\" (UniqueName: \"kubernetes.io/projected/0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec-kube-api-access-znxnb\") pod \"memcached-0\" (UID: \"0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec\") " pod="openstack/memcached-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.726863 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec\") " pod="openstack/memcached-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.726994 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec\") " pod="openstack/memcached-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.727128 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec-config-data\") pod \"memcached-0\" (UID: \"0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec\") " pod="openstack/memcached-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.829068 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znxnb\" (UniqueName: \"kubernetes.io/projected/0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec-kube-api-access-znxnb\") pod \"memcached-0\" (UID: \"0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec\") " pod="openstack/memcached-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.829158 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec\") " pod="openstack/memcached-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.829197 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec\") " pod="openstack/memcached-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.829235 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec-config-data\") pod \"memcached-0\" (UID: \"0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec\") " pod="openstack/memcached-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.829350 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec-kolla-config\") pod \"memcached-0\" (UID: \"0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec\") " pod="openstack/memcached-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.830129 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec-kolla-config\") pod \"memcached-0\" (UID: \"0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec\") " pod="openstack/memcached-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.832668 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec-config-data\") pod \"memcached-0\" (UID: \"0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec\") " pod="openstack/memcached-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.836057 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec-memcached-tls-certs\") pod \"memcached-0\" (UID: \"0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec\") " pod="openstack/memcached-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.841620 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec-combined-ca-bundle\") pod \"memcached-0\" (UID: \"0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec\") " pod="openstack/memcached-0" Dec 10 19:17:55 crc kubenswrapper[4828]: I1210 19:17:55.847651 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znxnb\" (UniqueName: \"kubernetes.io/projected/0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec-kube-api-access-znxnb\") pod \"memcached-0\" (UID: \"0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec\") " pod="openstack/memcached-0" Dec 10 19:17:56 crc kubenswrapper[4828]: I1210 19:17:56.005556 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 10 19:17:57 crc kubenswrapper[4828]: I1210 19:17:57.920320 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:17:57 crc kubenswrapper[4828]: I1210 19:17:57.925545 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 19:17:57 crc kubenswrapper[4828]: I1210 19:17:57.930729 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-dz5zn" Dec 10 19:17:57 crc kubenswrapper[4828]: I1210 19:17:57.958998 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:17:57 crc kubenswrapper[4828]: I1210 19:17:57.989840 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csdml\" (UniqueName: \"kubernetes.io/projected/7cd541f6-4cd3-42b6-9d59-ad4b6246044e-kube-api-access-csdml\") pod \"kube-state-metrics-0\" (UID: \"7cd541f6-4cd3-42b6-9d59-ad4b6246044e\") " pod="openstack/kube-state-metrics-0" Dec 10 19:17:58 crc kubenswrapper[4828]: I1210 19:17:58.092063 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csdml\" (UniqueName: \"kubernetes.io/projected/7cd541f6-4cd3-42b6-9d59-ad4b6246044e-kube-api-access-csdml\") pod \"kube-state-metrics-0\" (UID: \"7cd541f6-4cd3-42b6-9d59-ad4b6246044e\") " pod="openstack/kube-state-metrics-0" Dec 10 19:17:58 crc kubenswrapper[4828]: I1210 19:17:58.128972 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csdml\" (UniqueName: \"kubernetes.io/projected/7cd541f6-4cd3-42b6-9d59-ad4b6246044e-kube-api-access-csdml\") pod \"kube-state-metrics-0\" (UID: \"7cd541f6-4cd3-42b6-9d59-ad4b6246044e\") " pod="openstack/kube-state-metrics-0" Dec 10 19:17:58 crc kubenswrapper[4828]: I1210 19:17:58.259867 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 19:17:58 crc kubenswrapper[4828]: I1210 19:17:58.741036 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-nq29p"] Dec 10 19:17:58 crc kubenswrapper[4828]: I1210 19:17:58.742567 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-nq29p" Dec 10 19:17:58 crc kubenswrapper[4828]: I1210 19:17:58.750951 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-fzncj" Dec 10 19:17:58 crc kubenswrapper[4828]: I1210 19:17:58.751181 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Dec 10 19:17:58 crc kubenswrapper[4828]: I1210 19:17:58.759086 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-nq29p"] Dec 10 19:17:58 crc kubenswrapper[4828]: I1210 19:17:58.810350 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6624e745-c1ca-4dfd-9ee8-107310f3d368-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-nq29p\" (UID: \"6624e745-c1ca-4dfd-9ee8-107310f3d368\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-nq29p" Dec 10 19:17:58 crc kubenswrapper[4828]: I1210 19:17:58.810441 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcxd7\" (UniqueName: \"kubernetes.io/projected/6624e745-c1ca-4dfd-9ee8-107310f3d368-kube-api-access-pcxd7\") pod \"observability-ui-dashboards-7d5fb4cbfb-nq29p\" (UID: \"6624e745-c1ca-4dfd-9ee8-107310f3d368\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-nq29p" Dec 10 19:17:58 crc kubenswrapper[4828]: I1210 19:17:58.929189 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6624e745-c1ca-4dfd-9ee8-107310f3d368-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-nq29p\" (UID: \"6624e745-c1ca-4dfd-9ee8-107310f3d368\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-nq29p" Dec 10 19:17:58 crc kubenswrapper[4828]: I1210 19:17:58.929301 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcxd7\" (UniqueName: \"kubernetes.io/projected/6624e745-c1ca-4dfd-9ee8-107310f3d368-kube-api-access-pcxd7\") pod \"observability-ui-dashboards-7d5fb4cbfb-nq29p\" (UID: \"6624e745-c1ca-4dfd-9ee8-107310f3d368\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-nq29p" Dec 10 19:17:58 crc kubenswrapper[4828]: E1210 19:17:58.930382 4828 secret.go:188] Couldn't get secret openshift-operators/observability-ui-dashboards: secret "observability-ui-dashboards" not found Dec 10 19:17:58 crc kubenswrapper[4828]: E1210 19:17:58.930439 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6624e745-c1ca-4dfd-9ee8-107310f3d368-serving-cert podName:6624e745-c1ca-4dfd-9ee8-107310f3d368 nodeName:}" failed. No retries permitted until 2025-12-10 19:17:59.430420202 +0000 UTC m=+1359.941031207 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/6624e745-c1ca-4dfd-9ee8-107310f3d368-serving-cert") pod "observability-ui-dashboards-7d5fb4cbfb-nq29p" (UID: "6624e745-c1ca-4dfd-9ee8-107310f3d368") : secret "observability-ui-dashboards" not found Dec 10 19:17:58 crc kubenswrapper[4828]: I1210 19:17:58.985706 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcxd7\" (UniqueName: \"kubernetes.io/projected/6624e745-c1ca-4dfd-9ee8-107310f3d368-kube-api-access-pcxd7\") pod \"observability-ui-dashboards-7d5fb4cbfb-nq29p\" (UID: \"6624e745-c1ca-4dfd-9ee8-107310f3d368\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-nq29p" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.234755 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6946fcdcc7-xs7tg"] Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.238342 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.270840 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6946fcdcc7-xs7tg"] Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.327713 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.330052 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.338115 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.338499 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.338742 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.339503 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.346956 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.348502 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-jbqfm" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.351123 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/453bb320-a6c8-4de2-b79b-f77dcb42fe54-trusted-ca-bundle\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.351154 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/453bb320-a6c8-4de2-b79b-f77dcb42fe54-console-serving-cert\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.351185 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/453bb320-a6c8-4de2-b79b-f77dcb42fe54-service-ca\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.351217 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/453bb320-a6c8-4de2-b79b-f77dcb42fe54-console-oauth-config\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.351238 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m42p\" (UniqueName: \"kubernetes.io/projected/453bb320-a6c8-4de2-b79b-f77dcb42fe54-kube-api-access-8m42p\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.351317 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/453bb320-a6c8-4de2-b79b-f77dcb42fe54-oauth-serving-cert\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.351357 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/453bb320-a6c8-4de2-b79b-f77dcb42fe54-console-config\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.378185 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.454073 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.454118 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-config\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.454170 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/453bb320-a6c8-4de2-b79b-f77dcb42fe54-console-config\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.454212 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.454232 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.454260 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dppqj\" (UniqueName: \"kubernetes.io/projected/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-kube-api-access-dppqj\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.454890 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6624e745-c1ca-4dfd-9ee8-107310f3d368-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-nq29p\" (UID: \"6624e745-c1ca-4dfd-9ee8-107310f3d368\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-nq29p" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.454946 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/453bb320-a6c8-4de2-b79b-f77dcb42fe54-trusted-ca-bundle\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.454968 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/453bb320-a6c8-4de2-b79b-f77dcb42fe54-console-serving-cert\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.455024 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/453bb320-a6c8-4de2-b79b-f77dcb42fe54-service-ca\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.455054 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.455102 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/453bb320-a6c8-4de2-b79b-f77dcb42fe54-console-oauth-config\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.455143 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m42p\" (UniqueName: \"kubernetes.io/projected/453bb320-a6c8-4de2-b79b-f77dcb42fe54-kube-api-access-8m42p\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.455200 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.455314 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.455520 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/453bb320-a6c8-4de2-b79b-f77dcb42fe54-oauth-serving-cert\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.455638 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/453bb320-a6c8-4de2-b79b-f77dcb42fe54-console-config\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.457593 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/453bb320-a6c8-4de2-b79b-f77dcb42fe54-trusted-ca-bundle\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.457919 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/453bb320-a6c8-4de2-b79b-f77dcb42fe54-service-ca\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.458290 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/453bb320-a6c8-4de2-b79b-f77dcb42fe54-oauth-serving-cert\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.460590 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/453bb320-a6c8-4de2-b79b-f77dcb42fe54-console-oauth-config\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.460950 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/453bb320-a6c8-4de2-b79b-f77dcb42fe54-console-serving-cert\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.463264 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6624e745-c1ca-4dfd-9ee8-107310f3d368-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-nq29p\" (UID: \"6624e745-c1ca-4dfd-9ee8-107310f3d368\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-nq29p" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.475271 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m42p\" (UniqueName: \"kubernetes.io/projected/453bb320-a6c8-4de2-b79b-f77dcb42fe54-kube-api-access-8m42p\") pod \"console-6946fcdcc7-xs7tg\" (UID: \"453bb320-a6c8-4de2-b79b-f77dcb42fe54\") " pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.556814 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.556868 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.556900 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dppqj\" (UniqueName: \"kubernetes.io/projected/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-kube-api-access-dppqj\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.556936 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.556967 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.557010 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.557069 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.557097 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-config\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.557725 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.558018 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.561046 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-config\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.561101 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.561367 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.563757 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.564590 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.569238 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.574142 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dppqj\" (UniqueName: \"kubernetes.io/projected/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-kube-api-access-dppqj\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.597460 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.678208 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-nq29p" Dec 10 19:17:59 crc kubenswrapper[4828]: I1210 19:17:59.695893 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.569011 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.754109 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-krphm"] Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.755621 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.758487 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.758707 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-2bjff" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.760485 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.768850 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-krphm"] Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.791847 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-lldb4"] Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.795985 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.809810 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8fa35e66-ea0f-423d-b133-2170900e433f-scripts\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.810112 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8fa35e66-ea0f-423d-b133-2170900e433f-var-run\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.810231 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fa35e66-ea0f-423d-b133-2170900e433f-ovn-controller-tls-certs\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.810378 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa35e66-ea0f-423d-b133-2170900e433f-combined-ca-bundle\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.810501 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9pwl\" (UniqueName: \"kubernetes.io/projected/8fa35e66-ea0f-423d-b133-2170900e433f-kube-api-access-j9pwl\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.810645 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8fa35e66-ea0f-423d-b133-2170900e433f-var-log-ovn\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.810822 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8fa35e66-ea0f-423d-b133-2170900e433f-var-run-ovn\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.866094 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.878409 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-lldb4"] Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.878896 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.900216 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-7bljd" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.900490 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.900681 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.905227 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.905397 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.906889 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.916155 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8fa35e66-ea0f-423d-b133-2170900e433f-scripts\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.916244 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8fa35e66-ea0f-423d-b133-2170900e433f-var-run\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.916285 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fa35e66-ea0f-423d-b133-2170900e433f-ovn-controller-tls-certs\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.916313 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa35e66-ea0f-423d-b133-2170900e433f-combined-ca-bundle\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.916337 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9pwl\" (UniqueName: \"kubernetes.io/projected/8fa35e66-ea0f-423d-b133-2170900e433f-kube-api-access-j9pwl\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.916385 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8fa35e66-ea0f-423d-b133-2170900e433f-var-log-ovn\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.916428 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8fa35e66-ea0f-423d-b133-2170900e433f-var-run-ovn\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.918727 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8fa35e66-ea0f-423d-b133-2170900e433f-var-log-ovn\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.918917 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8fa35e66-ea0f-423d-b133-2170900e433f-var-run-ovn\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.922529 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8fa35e66-ea0f-423d-b133-2170900e433f-var-run\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.926718 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa35e66-ea0f-423d-b133-2170900e433f-combined-ca-bundle\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.942689 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8fa35e66-ea0f-423d-b133-2170900e433f-scripts\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.943732 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fa35e66-ea0f-423d-b133-2170900e433f-ovn-controller-tls-certs\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:01 crc kubenswrapper[4828]: I1210 19:18:01.946288 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9pwl\" (UniqueName: \"kubernetes.io/projected/8fa35e66-ea0f-423d-b133-2170900e433f-kube-api-access-j9pwl\") pod \"ovn-controller-krphm\" (UID: \"8fa35e66-ea0f-423d-b133-2170900e433f\") " pod="openstack/ovn-controller-krphm" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.018578 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.018816 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-var-run\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.018841 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7cdl\" (UniqueName: \"kubernetes.io/projected/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-kube-api-access-t7cdl\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.018876 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.018993 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-config\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.019012 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-var-log\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.019053 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-var-lib\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.019087 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.019127 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.019190 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-scripts\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.019213 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.019228 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.019249 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-etc-ovs\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.019289 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlrjt\" (UniqueName: \"kubernetes.io/projected/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-kube-api-access-hlrjt\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.085748 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-krphm" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.121020 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-var-lib\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.121162 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.121196 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.121305 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-scripts\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.121340 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.121355 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.121358 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-var-lib\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.121379 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-etc-ovs\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.121440 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlrjt\" (UniqueName: \"kubernetes.io/projected/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-kube-api-access-hlrjt\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.121533 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-etc-ovs\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.121563 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.121589 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.121877 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.122201 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-var-run\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.122312 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7cdl\" (UniqueName: \"kubernetes.io/projected/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-kube-api-access-t7cdl\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.122424 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.122521 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-config\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.122599 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-var-log\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.123222 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-var-run\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.123904 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-var-log\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.124007 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-scripts\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.124561 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-config\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.124563 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.124755 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.127246 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.133373 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.140525 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlrjt\" (UniqueName: \"kubernetes.io/projected/1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f-kube-api-access-hlrjt\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.142426 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7cdl\" (UniqueName: \"kubernetes.io/projected/517eff25-cf05-4abc-a46a-3d3dcd7e7f16-kube-api-access-t7cdl\") pod \"ovn-controller-ovs-lldb4\" (UID: \"517eff25-cf05-4abc-a46a-3d3dcd7e7f16\") " pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.148879 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f\") " pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.156859 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:02 crc kubenswrapper[4828]: I1210 19:18:02.206322 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.199778 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.206487 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.209450 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.210516 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.224511 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.227936 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-9jmhz" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.230276 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.381195 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/00c4966f-831b-40d7-a49f-da33b487cd84-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.381248 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00c4966f-831b-40d7-a49f-da33b487cd84-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.381499 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00c4966f-831b-40d7-a49f-da33b487cd84-config\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.381636 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.381685 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l98l\" (UniqueName: \"kubernetes.io/projected/00c4966f-831b-40d7-a49f-da33b487cd84-kube-api-access-6l98l\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.381714 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00c4966f-831b-40d7-a49f-da33b487cd84-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.381863 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/00c4966f-831b-40d7-a49f-da33b487cd84-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.381901 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/00c4966f-831b-40d7-a49f-da33b487cd84-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.483149 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/00c4966f-831b-40d7-a49f-da33b487cd84-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.483191 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00c4966f-831b-40d7-a49f-da33b487cd84-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.483251 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00c4966f-831b-40d7-a49f-da33b487cd84-config\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.483281 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.483297 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l98l\" (UniqueName: \"kubernetes.io/projected/00c4966f-831b-40d7-a49f-da33b487cd84-kube-api-access-6l98l\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.483315 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00c4966f-831b-40d7-a49f-da33b487cd84-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.483343 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/00c4966f-831b-40d7-a49f-da33b487cd84-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.483361 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/00c4966f-831b-40d7-a49f-da33b487cd84-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.483943 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.484152 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/00c4966f-831b-40d7-a49f-da33b487cd84-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.484831 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00c4966f-831b-40d7-a49f-da33b487cd84-config\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.484876 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00c4966f-831b-40d7-a49f-da33b487cd84-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.490350 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/00c4966f-831b-40d7-a49f-da33b487cd84-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.490501 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/00c4966f-831b-40d7-a49f-da33b487cd84-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.490968 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00c4966f-831b-40d7-a49f-da33b487cd84-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.508046 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l98l\" (UniqueName: \"kubernetes.io/projected/00c4966f-831b-40d7-a49f-da33b487cd84-kube-api-access-6l98l\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.510199 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-sb-0\" (UID: \"00c4966f-831b-40d7-a49f-da33b487cd84\") " pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:05 crc kubenswrapper[4828]: I1210 19:18:05.526057 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:08 crc kubenswrapper[4828]: W1210 19:18:08.462632 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddf0d133_f122_4bac_b746_222a86e29dd2.slice/crio-01ffbac41de359e8b1a3520a5b89ec1a05e643307f72a906dd990426fa47eb5b WatchSource:0}: Error finding container 01ffbac41de359e8b1a3520a5b89ec1a05e643307f72a906dd990426fa47eb5b: Status 404 returned error can't find the container with id 01ffbac41de359e8b1a3520a5b89ec1a05e643307f72a906dd990426fa47eb5b Dec 10 19:18:08 crc kubenswrapper[4828]: I1210 19:18:08.645642 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ddf0d133-f122-4bac-b746-222a86e29dd2","Type":"ContainerStarted","Data":"01ffbac41de359e8b1a3520a5b89ec1a05e643307f72a906dd990426fa47eb5b"} Dec 10 19:18:08 crc kubenswrapper[4828]: I1210 19:18:08.919933 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 10 19:18:13 crc kubenswrapper[4828]: I1210 19:18:13.434086 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-krphm"] Dec 10 19:18:13 crc kubenswrapper[4828]: I1210 19:18:13.703005 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec","Type":"ContainerStarted","Data":"06863041544ccff123f817abf323f8f6b26249ce1aeb691d8cd7d47bc82e9893"} Dec 10 19:18:13 crc kubenswrapper[4828]: W1210 19:18:13.907013 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fa35e66_ea0f_423d_b133_2170900e433f.slice/crio-d5d37dd782beb2aa9acfc6d0c62ab97b93b47116b28153486d3cb025c6a936d7 WatchSource:0}: Error finding container d5d37dd782beb2aa9acfc6d0c62ab97b93b47116b28153486d3cb025c6a936d7: Status 404 returned error can't find the container with id d5d37dd782beb2aa9acfc6d0c62ab97b93b47116b28153486d3cb025c6a936d7 Dec 10 19:18:13 crc kubenswrapper[4828]: E1210 19:18:13.931929 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 10 19:18:13 crc kubenswrapper[4828]: E1210 19:18:13.932464 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6rgjt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-znhwl_openstack(21b70614-6116-47aa-8336-afb75469d5ac): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:18:13 crc kubenswrapper[4828]: E1210 19:18:13.934223 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-znhwl" podUID="21b70614-6116-47aa-8336-afb75469d5ac" Dec 10 19:18:13 crc kubenswrapper[4828]: E1210 19:18:13.939282 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 10 19:18:13 crc kubenswrapper[4828]: E1210 19:18:13.940006 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6k6gf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-xkgl6_openstack(ebe57772-c202-4d69-986f-aa63f6eeed94): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:18:13 crc kubenswrapper[4828]: E1210 19:18:13.941436 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5ccc8479f9-xkgl6" podUID="ebe57772-c202-4d69-986f-aa63f6eeed94" Dec 10 19:18:13 crc kubenswrapper[4828]: E1210 19:18:13.941677 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 10 19:18:13 crc kubenswrapper[4828]: E1210 19:18:13.941970 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xrmhx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-2wdrq_openstack(cb237099-ff45-4060-bb7d-d506f069837f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:18:13 crc kubenswrapper[4828]: E1210 19:18:13.943184 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-2wdrq" podUID="cb237099-ff45-4060-bb7d-d506f069837f" Dec 10 19:18:14 crc kubenswrapper[4828]: I1210 19:18:14.395043 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-nq29p"] Dec 10 19:18:14 crc kubenswrapper[4828]: W1210 19:18:14.403556 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6624e745_c1ca_4dfd_9ee8_107310f3d368.slice/crio-d4beeb75363fbdad1020807421b9512927ff3d5feb0a2a3a9d37f5a347336609 WatchSource:0}: Error finding container d4beeb75363fbdad1020807421b9512927ff3d5feb0a2a3a9d37f5a347336609: Status 404 returned error can't find the container with id d4beeb75363fbdad1020807421b9512927ff3d5feb0a2a3a9d37f5a347336609 Dec 10 19:18:14 crc kubenswrapper[4828]: I1210 19:18:14.714264 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-nq29p" event={"ID":"6624e745-c1ca-4dfd-9ee8-107310f3d368","Type":"ContainerStarted","Data":"d4beeb75363fbdad1020807421b9512927ff3d5feb0a2a3a9d37f5a347336609"} Dec 10 19:18:14 crc kubenswrapper[4828]: I1210 19:18:14.718509 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-krphm" event={"ID":"8fa35e66-ea0f-423d-b133-2170900e433f","Type":"ContainerStarted","Data":"d5d37dd782beb2aa9acfc6d0c62ab97b93b47116b28153486d3cb025c6a936d7"} Dec 10 19:18:14 crc kubenswrapper[4828]: I1210 19:18:14.854006 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:18:14 crc kubenswrapper[4828]: I1210 19:18:14.887480 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:18:14 crc kubenswrapper[4828]: I1210 19:18:14.917746 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 10 19:18:14 crc kubenswrapper[4828]: W1210 19:18:14.942264 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b0f0962_54b6_46c3_8c5d_d008874f754f.slice/crio-13e037dd318ad231d752d689c0cff17fa456d6e3107b25fed66551ba375944bc WatchSource:0}: Error finding container 13e037dd318ad231d752d689c0cff17fa456d6e3107b25fed66551ba375944bc: Status 404 returned error can't find the container with id 13e037dd318ad231d752d689c0cff17fa456d6e3107b25fed66551ba375944bc Dec 10 19:18:15 crc kubenswrapper[4828]: E1210 19:18:15.009493 4828 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 10 19:18:15 crc kubenswrapper[4828]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/ebe57772-c202-4d69-986f-aa63f6eeed94/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 10 19:18:15 crc kubenswrapper[4828]: > podSandboxID="aa2d2d10ea00fb51152a67173715a9da8f05237af7d980edcd00235ea39da53a" Dec 10 19:18:15 crc kubenswrapper[4828]: E1210 19:18:15.009694 4828 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 10 19:18:15 crc kubenswrapper[4828]: init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6k6gf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-xkgl6_openstack(ebe57772-c202-4d69-986f-aa63f6eeed94): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/ebe57772-c202-4d69-986f-aa63f6eeed94/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 10 19:18:15 crc kubenswrapper[4828]: > logger="UnhandledError" Dec 10 19:18:15 crc kubenswrapper[4828]: E1210 19:18:15.011205 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/ebe57772-c202-4d69-986f-aa63f6eeed94/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-5ccc8479f9-xkgl6" podUID="ebe57772-c202-4d69-986f-aa63f6eeed94" Dec 10 19:18:15 crc kubenswrapper[4828]: I1210 19:18:15.077550 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-lldb4"] Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:15.436696 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6946fcdcc7-xs7tg"] Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:15.528955 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:15.754490 4828 generic.go:334] "Generic (PLEG): container finished" podID="d2630e12-86b3-45d3-84c0-dd50f5cf8c15" containerID="c4ad2ff657869f590f022f32220e87ccadf7b672abfe55db3e29bd7626616c1a" exitCode=0 Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:15.754561 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" event={"ID":"d2630e12-86b3-45d3-84c0-dd50f5cf8c15","Type":"ContainerDied","Data":"c4ad2ff657869f590f022f32220e87ccadf7b672abfe55db3e29bd7626616c1a"} Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:15.774403 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-2wdrq" event={"ID":"cb237099-ff45-4060-bb7d-d506f069837f","Type":"ContainerDied","Data":"bdee8b24f9ae37b0a37ee11ff65e6fbda928f19182c9b09263fa51b58779ece0"} Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:15.774437 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdee8b24f9ae37b0a37ee11ff65e6fbda928f19182c9b09263fa51b58779ece0" Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:15.775665 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0","Type":"ContainerStarted","Data":"c8e245b3412b3b73e4489df2bdc8ee6671ab3b7fa58a6b1fd127953972800fed"} Dec 10 19:18:16 crc kubenswrapper[4828]: W1210 19:18:15.779102 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00c4966f_831b_40d7_a49f_da33b487cd84.slice/crio-7d37ee073900349cc3decd7c4eadea71480da6be69172c8475b4d9517071c6f1 WatchSource:0}: Error finding container 7d37ee073900349cc3decd7c4eadea71480da6be69172c8475b4d9517071c6f1: Status 404 returned error can't find the container with id 7d37ee073900349cc3decd7c4eadea71480da6be69172c8475b4d9517071c6f1 Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:15.810327 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7cd541f6-4cd3-42b6-9d59-ad4b6246044e","Type":"ContainerStarted","Data":"31d7712b7667da8678af7c12d0e60c79cc9d8779cfd4545f4f0adb5aeb896450"} Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:15.810439 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8b0f0962-54b6-46c3-8c5d-d008874f754f","Type":"ContainerStarted","Data":"13e037dd318ad231d752d689c0cff17fa456d6e3107b25fed66551ba375944bc"} Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:15.810456 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lldb4" event={"ID":"517eff25-cf05-4abc-a46a-3d3dcd7e7f16","Type":"ContainerStarted","Data":"f251d566af975478d8fb5dc7940879597919aa38075ebe6f01dfd881a188dfbd"} Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:15.819152 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6946fcdcc7-xs7tg" event={"ID":"453bb320-a6c8-4de2-b79b-f77dcb42fe54","Type":"ContainerStarted","Data":"1a97e4b6e5a3fa61bee0900579a4e05bd6e0b190810f39cd9558cdfd4c80831f"} Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.141623 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.143094 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2wdrq" Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.157989 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-znhwl" Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.323487 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb237099-ff45-4060-bb7d-d506f069837f-dns-svc\") pod \"cb237099-ff45-4060-bb7d-d506f069837f\" (UID: \"cb237099-ff45-4060-bb7d-d506f069837f\") " Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.323558 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrmhx\" (UniqueName: \"kubernetes.io/projected/cb237099-ff45-4060-bb7d-d506f069837f-kube-api-access-xrmhx\") pod \"cb237099-ff45-4060-bb7d-d506f069837f\" (UID: \"cb237099-ff45-4060-bb7d-d506f069837f\") " Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.323656 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb237099-ff45-4060-bb7d-d506f069837f-config\") pod \"cb237099-ff45-4060-bb7d-d506f069837f\" (UID: \"cb237099-ff45-4060-bb7d-d506f069837f\") " Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.323689 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21b70614-6116-47aa-8336-afb75469d5ac-config\") pod \"21b70614-6116-47aa-8336-afb75469d5ac\" (UID: \"21b70614-6116-47aa-8336-afb75469d5ac\") " Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.323756 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rgjt\" (UniqueName: \"kubernetes.io/projected/21b70614-6116-47aa-8336-afb75469d5ac-kube-api-access-6rgjt\") pod \"21b70614-6116-47aa-8336-afb75469d5ac\" (UID: \"21b70614-6116-47aa-8336-afb75469d5ac\") " Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.324177 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb237099-ff45-4060-bb7d-d506f069837f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cb237099-ff45-4060-bb7d-d506f069837f" (UID: "cb237099-ff45-4060-bb7d-d506f069837f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.324185 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb237099-ff45-4060-bb7d-d506f069837f-config" (OuterVolumeSpecName: "config") pod "cb237099-ff45-4060-bb7d-d506f069837f" (UID: "cb237099-ff45-4060-bb7d-d506f069837f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.324337 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21b70614-6116-47aa-8336-afb75469d5ac-config" (OuterVolumeSpecName: "config") pod "21b70614-6116-47aa-8336-afb75469d5ac" (UID: "21b70614-6116-47aa-8336-afb75469d5ac"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.333096 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb237099-ff45-4060-bb7d-d506f069837f-kube-api-access-xrmhx" (OuterVolumeSpecName: "kube-api-access-xrmhx") pod "cb237099-ff45-4060-bb7d-d506f069837f" (UID: "cb237099-ff45-4060-bb7d-d506f069837f"). InnerVolumeSpecName "kube-api-access-xrmhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.338324 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21b70614-6116-47aa-8336-afb75469d5ac-kube-api-access-6rgjt" (OuterVolumeSpecName: "kube-api-access-6rgjt") pod "21b70614-6116-47aa-8336-afb75469d5ac" (UID: "21b70614-6116-47aa-8336-afb75469d5ac"). InnerVolumeSpecName "kube-api-access-6rgjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.425702 4828 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cb237099-ff45-4060-bb7d-d506f069837f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.425732 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrmhx\" (UniqueName: \"kubernetes.io/projected/cb237099-ff45-4060-bb7d-d506f069837f-kube-api-access-xrmhx\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.425743 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb237099-ff45-4060-bb7d-d506f069837f-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.425751 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21b70614-6116-47aa-8336-afb75469d5ac-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.425763 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rgjt\" (UniqueName: \"kubernetes.io/projected/21b70614-6116-47aa-8336-afb75469d5ac-kube-api-access-6rgjt\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.827305 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-znhwl" event={"ID":"21b70614-6116-47aa-8336-afb75469d5ac","Type":"ContainerDied","Data":"011300d0f201745e8d832733121e9bbd3c99815c3a89705b93af0e226ac0a6ad"} Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.827434 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-znhwl" Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.835918 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"00c4966f-831b-40d7-a49f-da33b487cd84","Type":"ContainerStarted","Data":"7d37ee073900349cc3decd7c4eadea71480da6be69172c8475b4d9517071c6f1"} Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.837214 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6946fcdcc7-xs7tg" event={"ID":"453bb320-a6c8-4de2-b79b-f77dcb42fe54","Type":"ContainerStarted","Data":"6dcfb2aacc6081935cc250562fc5994cbd24cd3de885c42d24f4e8e309221fd1"} Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.840671 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b24f93de-d604-499b-be8e-2e934d119192","Type":"ContainerStarted","Data":"5505417dc4fa51ffbfa46576fe4ceb89ae8a05311568421b1979522bbc329385"} Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.845469 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" event={"ID":"d2630e12-86b3-45d3-84c0-dd50f5cf8c15","Type":"ContainerStarted","Data":"a09f7f4c7627b83fbdb3841a2f8006809ec93cb5fda851ad0dbf804255d998eb"} Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.845636 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.848201 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-2wdrq" Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.848233 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fd6f7666-6cc8-478d-9acf-21cfbad541c9","Type":"ContainerStarted","Data":"9fb4907035796b871225aeb7750d239070e57fcc3199b135138f56a1c472ce23"} Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.866824 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6946fcdcc7-xs7tg" podStartSLOduration=17.866790522 podStartE2EDuration="17.866790522s" podCreationTimestamp="2025-12-10 19:17:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:16.855675652 +0000 UTC m=+1377.366286667" watchObservedRunningTime="2025-12-10 19:18:16.866790522 +0000 UTC m=+1377.377401527" Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.912057 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-znhwl"] Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.937258 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-znhwl"] Dec 10 19:18:16 crc kubenswrapper[4828]: I1210 19:18:16.976678 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" podStartSLOduration=3.828457317 podStartE2EDuration="25.976660502s" podCreationTimestamp="2025-12-10 19:17:51 +0000 UTC" firstStartedPulling="2025-12-10 19:17:52.097868991 +0000 UTC m=+1352.608479996" lastFinishedPulling="2025-12-10 19:18:14.246072176 +0000 UTC m=+1374.756683181" observedRunningTime="2025-12-10 19:18:16.964430321 +0000 UTC m=+1377.475041336" watchObservedRunningTime="2025-12-10 19:18:16.976660502 +0000 UTC m=+1377.487271507" Dec 10 19:18:17 crc kubenswrapper[4828]: I1210 19:18:17.011784 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2wdrq"] Dec 10 19:18:17 crc kubenswrapper[4828]: I1210 19:18:17.018570 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-2wdrq"] Dec 10 19:18:17 crc kubenswrapper[4828]: I1210 19:18:17.801211 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21b70614-6116-47aa-8336-afb75469d5ac" path="/var/lib/kubelet/pods/21b70614-6116-47aa-8336-afb75469d5ac/volumes" Dec 10 19:18:17 crc kubenswrapper[4828]: I1210 19:18:17.801925 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb237099-ff45-4060-bb7d-d506f069837f" path="/var/lib/kubelet/pods/cb237099-ff45-4060-bb7d-d506f069837f/volumes" Dec 10 19:18:18 crc kubenswrapper[4828]: W1210 19:18:18.251673 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bc5f1a9_6af4_419b_a3b9_6f41b6113e8f.slice/crio-f2f37cb7d8d6a1e76574ecb20637f167f536d39df4123cde26638eaf25331aa5 WatchSource:0}: Error finding container f2f37cb7d8d6a1e76574ecb20637f167f536d39df4123cde26638eaf25331aa5: Status 404 returned error can't find the container with id f2f37cb7d8d6a1e76574ecb20637f167f536d39df4123cde26638eaf25331aa5 Dec 10 19:18:18 crc kubenswrapper[4828]: I1210 19:18:18.868287 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f","Type":"ContainerStarted","Data":"f2f37cb7d8d6a1e76574ecb20637f167f536d39df4123cde26638eaf25331aa5"} Dec 10 19:18:19 crc kubenswrapper[4828]: I1210 19:18:19.569915 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:18:19 crc kubenswrapper[4828]: I1210 19:18:19.569979 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:18:19 crc kubenswrapper[4828]: I1210 19:18:19.574952 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:18:19 crc kubenswrapper[4828]: I1210 19:18:19.882330 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6946fcdcc7-xs7tg" Dec 10 19:18:20 crc kubenswrapper[4828]: I1210 19:18:20.040463 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-7994d84d8-5xz2j"] Dec 10 19:18:21 crc kubenswrapper[4828]: I1210 19:18:21.461003 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" Dec 10 19:18:21 crc kubenswrapper[4828]: I1210 19:18:21.521271 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-xkgl6"] Dec 10 19:18:22 crc kubenswrapper[4828]: I1210 19:18:22.488567 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-xkgl6" Dec 10 19:18:22 crc kubenswrapper[4828]: I1210 19:18:22.586402 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6k6gf\" (UniqueName: \"kubernetes.io/projected/ebe57772-c202-4d69-986f-aa63f6eeed94-kube-api-access-6k6gf\") pod \"ebe57772-c202-4d69-986f-aa63f6eeed94\" (UID: \"ebe57772-c202-4d69-986f-aa63f6eeed94\") " Dec 10 19:18:22 crc kubenswrapper[4828]: I1210 19:18:22.586728 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebe57772-c202-4d69-986f-aa63f6eeed94-dns-svc\") pod \"ebe57772-c202-4d69-986f-aa63f6eeed94\" (UID: \"ebe57772-c202-4d69-986f-aa63f6eeed94\") " Dec 10 19:18:22 crc kubenswrapper[4828]: I1210 19:18:22.586846 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebe57772-c202-4d69-986f-aa63f6eeed94-config\") pod \"ebe57772-c202-4d69-986f-aa63f6eeed94\" (UID: \"ebe57772-c202-4d69-986f-aa63f6eeed94\") " Dec 10 19:18:22 crc kubenswrapper[4828]: I1210 19:18:22.590353 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebe57772-c202-4d69-986f-aa63f6eeed94-kube-api-access-6k6gf" (OuterVolumeSpecName: "kube-api-access-6k6gf") pod "ebe57772-c202-4d69-986f-aa63f6eeed94" (UID: "ebe57772-c202-4d69-986f-aa63f6eeed94"). InnerVolumeSpecName "kube-api-access-6k6gf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:22 crc kubenswrapper[4828]: I1210 19:18:22.621054 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebe57772-c202-4d69-986f-aa63f6eeed94-config" (OuterVolumeSpecName: "config") pod "ebe57772-c202-4d69-986f-aa63f6eeed94" (UID: "ebe57772-c202-4d69-986f-aa63f6eeed94"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:22 crc kubenswrapper[4828]: I1210 19:18:22.622235 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebe57772-c202-4d69-986f-aa63f6eeed94-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ebe57772-c202-4d69-986f-aa63f6eeed94" (UID: "ebe57772-c202-4d69-986f-aa63f6eeed94"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:22 crc kubenswrapper[4828]: I1210 19:18:22.689927 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6k6gf\" (UniqueName: \"kubernetes.io/projected/ebe57772-c202-4d69-986f-aa63f6eeed94-kube-api-access-6k6gf\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:22 crc kubenswrapper[4828]: I1210 19:18:22.689989 4828 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebe57772-c202-4d69-986f-aa63f6eeed94-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:22 crc kubenswrapper[4828]: I1210 19:18:22.690003 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebe57772-c202-4d69-986f-aa63f6eeed94-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:22 crc kubenswrapper[4828]: I1210 19:18:22.930116 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-xkgl6" event={"ID":"ebe57772-c202-4d69-986f-aa63f6eeed94","Type":"ContainerDied","Data":"aa2d2d10ea00fb51152a67173715a9da8f05237af7d980edcd00235ea39da53a"} Dec 10 19:18:22 crc kubenswrapper[4828]: I1210 19:18:22.930171 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-xkgl6" Dec 10 19:18:22 crc kubenswrapper[4828]: I1210 19:18:22.997784 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-xkgl6"] Dec 10 19:18:23 crc kubenswrapper[4828]: I1210 19:18:23.006271 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-xkgl6"] Dec 10 19:18:23 crc kubenswrapper[4828]: I1210 19:18:23.808272 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebe57772-c202-4d69-986f-aa63f6eeed94" path="/var/lib/kubelet/pods/ebe57772-c202-4d69-986f-aa63f6eeed94/volumes" Dec 10 19:18:25 crc kubenswrapper[4828]: I1210 19:18:25.960703 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec","Type":"ContainerStarted","Data":"8d411d1e86fec0e1f1331aaa20340c03148d4d97831bfeedb9a7f11d15781e7b"} Dec 10 19:18:25 crc kubenswrapper[4828]: I1210 19:18:25.961312 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 10 19:18:25 crc kubenswrapper[4828]: I1210 19:18:25.964960 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ddf0d133-f122-4bac-b746-222a86e29dd2","Type":"ContainerStarted","Data":"64dc70276444f29bf4d25d1cf2f99554fa7c03b6af52c2a42f430080fb087bf4"} Dec 10 19:18:25 crc kubenswrapper[4828]: I1210 19:18:25.970523 4828 generic.go:334] "Generic (PLEG): container finished" podID="517eff25-cf05-4abc-a46a-3d3dcd7e7f16" containerID="37a690f4be0cf99665f9e69ecab1936f27e38778135827273f38740b5259cbb7" exitCode=0 Dec 10 19:18:25 crc kubenswrapper[4828]: I1210 19:18:25.970619 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lldb4" event={"ID":"517eff25-cf05-4abc-a46a-3d3dcd7e7f16","Type":"ContainerDied","Data":"37a690f4be0cf99665f9e69ecab1936f27e38778135827273f38740b5259cbb7"} Dec 10 19:18:25 crc kubenswrapper[4828]: I1210 19:18:25.974321 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-nq29p" event={"ID":"6624e745-c1ca-4dfd-9ee8-107310f3d368","Type":"ContainerStarted","Data":"267abe8e5d7f56a095c0c396619ced5b378baab7430cb7ef55d38f4ed3d39e9e"} Dec 10 19:18:25 crc kubenswrapper[4828]: I1210 19:18:25.982184 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=20.752174224 podStartE2EDuration="30.982136005s" podCreationTimestamp="2025-12-10 19:17:55 +0000 UTC" firstStartedPulling="2025-12-10 19:18:13.038596138 +0000 UTC m=+1373.549207143" lastFinishedPulling="2025-12-10 19:18:23.268557919 +0000 UTC m=+1383.779168924" observedRunningTime="2025-12-10 19:18:25.982035703 +0000 UTC m=+1386.492646738" watchObservedRunningTime="2025-12-10 19:18:25.982136005 +0000 UTC m=+1386.492747010" Dec 10 19:18:25 crc kubenswrapper[4828]: I1210 19:18:25.983272 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f","Type":"ContainerStarted","Data":"a905d12a8983b627fa7741e74d42654660ae1085c21c6dcb3684309b380c9035"} Dec 10 19:18:25 crc kubenswrapper[4828]: I1210 19:18:25.988209 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"00c4966f-831b-40d7-a49f-da33b487cd84","Type":"ContainerStarted","Data":"327a78e15c69cb2aee4e2af334c9090b90c6c079ffaa2435919c51c843aa1da8"} Dec 10 19:18:25 crc kubenswrapper[4828]: I1210 19:18:25.991005 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-krphm" event={"ID":"8fa35e66-ea0f-423d-b133-2170900e433f","Type":"ContainerStarted","Data":"19a3397acc2e123a971ee735250b7267d4fa0bf584d63573d3bc26ced71e882d"} Dec 10 19:18:25 crc kubenswrapper[4828]: I1210 19:18:25.992546 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-krphm" Dec 10 19:18:25 crc kubenswrapper[4828]: I1210 19:18:25.997160 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7cd541f6-4cd3-42b6-9d59-ad4b6246044e","Type":"ContainerStarted","Data":"0e61cbc0b1b55deb42c336115cad8dfaa9129689804f341614c3edf3541b1cda"} Dec 10 19:18:25 crc kubenswrapper[4828]: I1210 19:18:25.997312 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 10 19:18:25 crc kubenswrapper[4828]: I1210 19:18:25.999336 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8b0f0962-54b6-46c3-8c5d-d008874f754f","Type":"ContainerStarted","Data":"9054ca14867429ac3ebbf0ed25914ee23c67662f0ce6b1d464e479651c98ee14"} Dec 10 19:18:26 crc kubenswrapper[4828]: I1210 19:18:26.027652 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-nq29p" podStartSLOduration=19.405480302 podStartE2EDuration="28.027632215s" podCreationTimestamp="2025-12-10 19:17:58 +0000 UTC" firstStartedPulling="2025-12-10 19:18:14.409297927 +0000 UTC m=+1374.919908932" lastFinishedPulling="2025-12-10 19:18:23.03144984 +0000 UTC m=+1383.542060845" observedRunningTime="2025-12-10 19:18:26.018379205 +0000 UTC m=+1386.528990230" watchObservedRunningTime="2025-12-10 19:18:26.027632215 +0000 UTC m=+1386.538243220" Dec 10 19:18:26 crc kubenswrapper[4828]: I1210 19:18:26.074972 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=19.497803717 podStartE2EDuration="29.074950694s" podCreationTimestamp="2025-12-10 19:17:57 +0000 UTC" firstStartedPulling="2025-12-10 19:18:14.929774855 +0000 UTC m=+1375.440385860" lastFinishedPulling="2025-12-10 19:18:24.506921832 +0000 UTC m=+1385.017532837" observedRunningTime="2025-12-10 19:18:26.060334489 +0000 UTC m=+1386.570945524" watchObservedRunningTime="2025-12-10 19:18:26.074950694 +0000 UTC m=+1386.585561699" Dec 10 19:18:26 crc kubenswrapper[4828]: I1210 19:18:26.112336 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-krphm" podStartSLOduration=15.45273275 podStartE2EDuration="25.112316924s" podCreationTimestamp="2025-12-10 19:18:01 +0000 UTC" firstStartedPulling="2025-12-10 19:18:13.915339346 +0000 UTC m=+1374.425950351" lastFinishedPulling="2025-12-10 19:18:23.57492352 +0000 UTC m=+1384.085534525" observedRunningTime="2025-12-10 19:18:26.105934071 +0000 UTC m=+1386.616545076" watchObservedRunningTime="2025-12-10 19:18:26.112316924 +0000 UTC m=+1386.622927929" Dec 10 19:18:27 crc kubenswrapper[4828]: I1210 19:18:27.027588 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lldb4" event={"ID":"517eff25-cf05-4abc-a46a-3d3dcd7e7f16","Type":"ContainerStarted","Data":"480965bd7074a2f31ea44120e1cb75122a964edd9d13d731dad7e1b3b5a72459"} Dec 10 19:18:27 crc kubenswrapper[4828]: I1210 19:18:27.028097 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-lldb4" event={"ID":"517eff25-cf05-4abc-a46a-3d3dcd7e7f16","Type":"ContainerStarted","Data":"00083ce5ed90d46c855631bfad5a47b213f7841d2609f35944a234faf91ac97b"} Dec 10 19:18:27 crc kubenswrapper[4828]: I1210 19:18:27.029162 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:27 crc kubenswrapper[4828]: I1210 19:18:27.029655 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:27 crc kubenswrapper[4828]: I1210 19:18:27.060358 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-lldb4" podStartSLOduration=17.887237124 podStartE2EDuration="26.060340429s" podCreationTimestamp="2025-12-10 19:18:01 +0000 UTC" firstStartedPulling="2025-12-10 19:18:15.095583597 +0000 UTC m=+1375.606194602" lastFinishedPulling="2025-12-10 19:18:23.268686912 +0000 UTC m=+1383.779297907" observedRunningTime="2025-12-10 19:18:27.0540764 +0000 UTC m=+1387.564687405" watchObservedRunningTime="2025-12-10 19:18:27.060340429 +0000 UTC m=+1387.570951434" Dec 10 19:18:28 crc kubenswrapper[4828]: I1210 19:18:28.039904 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0","Type":"ContainerStarted","Data":"1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3"} Dec 10 19:18:29 crc kubenswrapper[4828]: I1210 19:18:29.050419 4828 generic.go:334] "Generic (PLEG): container finished" podID="8b0f0962-54b6-46c3-8c5d-d008874f754f" containerID="9054ca14867429ac3ebbf0ed25914ee23c67662f0ce6b1d464e479651c98ee14" exitCode=0 Dec 10 19:18:29 crc kubenswrapper[4828]: I1210 19:18:29.050515 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8b0f0962-54b6-46c3-8c5d-d008874f754f","Type":"ContainerDied","Data":"9054ca14867429ac3ebbf0ed25914ee23c67662f0ce6b1d464e479651c98ee14"} Dec 10 19:18:29 crc kubenswrapper[4828]: I1210 19:18:29.053622 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"00c4966f-831b-40d7-a49f-da33b487cd84","Type":"ContainerStarted","Data":"b0380727eeb6bd924e37fd09fe47d9ab3291d5953514ed8c8b53bddccf632c7c"} Dec 10 19:18:29 crc kubenswrapper[4828]: I1210 19:18:29.058911 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f","Type":"ContainerStarted","Data":"12ad003ea19c80e341b78b205ade22dfb15346a9b88c780d27b9d291deccca6c"} Dec 10 19:18:29 crc kubenswrapper[4828]: I1210 19:18:29.110388 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=12.443989043 podStartE2EDuration="25.11036969s" podCreationTimestamp="2025-12-10 19:18:04 +0000 UTC" firstStartedPulling="2025-12-10 19:18:15.807106119 +0000 UTC m=+1376.317717124" lastFinishedPulling="2025-12-10 19:18:28.473486756 +0000 UTC m=+1388.984097771" observedRunningTime="2025-12-10 19:18:29.104361517 +0000 UTC m=+1389.614972522" watchObservedRunningTime="2025-12-10 19:18:29.11036969 +0000 UTC m=+1389.620980695" Dec 10 19:18:29 crc kubenswrapper[4828]: I1210 19:18:29.139428 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=18.932752653 podStartE2EDuration="29.139410525s" podCreationTimestamp="2025-12-10 19:18:00 +0000 UTC" firstStartedPulling="2025-12-10 19:18:18.254213563 +0000 UTC m=+1378.764824578" lastFinishedPulling="2025-12-10 19:18:28.460871445 +0000 UTC m=+1388.971482450" observedRunningTime="2025-12-10 19:18:29.124785179 +0000 UTC m=+1389.635396184" watchObservedRunningTime="2025-12-10 19:18:29.139410525 +0000 UTC m=+1389.650021530" Dec 10 19:18:29 crc kubenswrapper[4828]: I1210 19:18:29.207490 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:29 crc kubenswrapper[4828]: I1210 19:18:29.251205 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:29 crc kubenswrapper[4828]: I1210 19:18:29.526807 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:29 crc kubenswrapper[4828]: I1210 19:18:29.567652 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.068110 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8b0f0962-54b6-46c3-8c5d-d008874f754f","Type":"ContainerStarted","Data":"3b06826c8932d5d1fba8ab6d23118435a4f97ad76be82d8c57554cef4d81a13a"} Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.069771 4828 generic.go:334] "Generic (PLEG): container finished" podID="ddf0d133-f122-4bac-b746-222a86e29dd2" containerID="64dc70276444f29bf4d25d1cf2f99554fa7c03b6af52c2a42f430080fb087bf4" exitCode=0 Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.069896 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ddf0d133-f122-4bac-b746-222a86e29dd2","Type":"ContainerDied","Data":"64dc70276444f29bf4d25d1cf2f99554fa7c03b6af52c2a42f430080fb087bf4"} Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.070108 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.070491 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.101113 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=27.569523544 podStartE2EDuration="36.101092909s" podCreationTimestamp="2025-12-10 19:17:54 +0000 UTC" firstStartedPulling="2025-12-10 19:18:14.948708907 +0000 UTC m=+1375.459319912" lastFinishedPulling="2025-12-10 19:18:23.480278272 +0000 UTC m=+1383.990889277" observedRunningTime="2025-12-10 19:18:30.096733611 +0000 UTC m=+1390.607344626" watchObservedRunningTime="2025-12-10 19:18:30.101092909 +0000 UTC m=+1390.611703914" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.123549 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.143021 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.369781 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-8hxrs"] Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.389685 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.396324 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.406523 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-8hxrs"] Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.461959 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-8hxrs\" (UID: \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\") " pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.462037 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-8hxrs\" (UID: \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\") " pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.462180 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-config\") pod \"dnsmasq-dns-6bc7876d45-8hxrs\" (UID: \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\") " pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.462212 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qshm\" (UniqueName: \"kubernetes.io/projected/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-kube-api-access-6qshm\") pod \"dnsmasq-dns-6bc7876d45-8hxrs\" (UID: \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\") " pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.522589 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-zmqsz"] Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.536950 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.540418 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.564376 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-8hxrs\" (UID: \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\") " pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.564478 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-config\") pod \"dnsmasq-dns-6bc7876d45-8hxrs\" (UID: \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\") " pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.564507 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qshm\" (UniqueName: \"kubernetes.io/projected/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-kube-api-access-6qshm\") pod \"dnsmasq-dns-6bc7876d45-8hxrs\" (UID: \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\") " pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.564562 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-8hxrs\" (UID: \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\") " pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.565316 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-8hxrs\" (UID: \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\") " pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.565835 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-8hxrs\" (UID: \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\") " pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.566081 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-zmqsz"] Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.566288 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-config\") pod \"dnsmasq-dns-6bc7876d45-8hxrs\" (UID: \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\") " pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.618658 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qshm\" (UniqueName: \"kubernetes.io/projected/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-kube-api-access-6qshm\") pod \"dnsmasq-dns-6bc7876d45-8hxrs\" (UID: \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\") " pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.642289 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.650746 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.653874 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.654072 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-mzzg9" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.654248 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.669080 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5204157-80d9-4397-83a5-865cf0dd293f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.669223 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99gr2\" (UniqueName: \"kubernetes.io/projected/b5204157-80d9-4397-83a5-865cf0dd293f-kube-api-access-99gr2\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.669269 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5204157-80d9-4397-83a5-865cf0dd293f-combined-ca-bundle\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.669288 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b5204157-80d9-4397-83a5-865cf0dd293f-ovs-rundir\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.669314 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5204157-80d9-4397-83a5-865cf0dd293f-config\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.669398 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b5204157-80d9-4397-83a5-865cf0dd293f-ovn-rundir\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.669621 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.706962 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-8hxrs"] Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.707765 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.768856 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.773179 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c323eb01-f638-4ec8-98a4-711fbd9063f2-scripts\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.773267 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c323eb01-f638-4ec8-98a4-711fbd9063f2-config\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.773293 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtld5\" (UniqueName: \"kubernetes.io/projected/c323eb01-f638-4ec8-98a4-711fbd9063f2-kube-api-access-qtld5\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.773363 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c323eb01-f638-4ec8-98a4-711fbd9063f2-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.773399 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99gr2\" (UniqueName: \"kubernetes.io/projected/b5204157-80d9-4397-83a5-865cf0dd293f-kube-api-access-99gr2\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.773440 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5204157-80d9-4397-83a5-865cf0dd293f-combined-ca-bundle\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.773462 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b5204157-80d9-4397-83a5-865cf0dd293f-ovs-rundir\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.773492 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5204157-80d9-4397-83a5-865cf0dd293f-config\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.773511 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c323eb01-f638-4ec8-98a4-711fbd9063f2-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.773529 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c323eb01-f638-4ec8-98a4-711fbd9063f2-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.773552 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b5204157-80d9-4397-83a5-865cf0dd293f-ovn-rundir\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.773607 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5204157-80d9-4397-83a5-865cf0dd293f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.773637 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c323eb01-f638-4ec8-98a4-711fbd9063f2-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.774392 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/b5204157-80d9-4397-83a5-865cf0dd293f-ovn-rundir\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.774392 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/b5204157-80d9-4397-83a5-865cf0dd293f-ovs-rundir\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.775139 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5204157-80d9-4397-83a5-865cf0dd293f-config\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.781788 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5204157-80d9-4397-83a5-865cf0dd293f-combined-ca-bundle\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.784405 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5204157-80d9-4397-83a5-865cf0dd293f-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.823382 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99gr2\" (UniqueName: \"kubernetes.io/projected/b5204157-80d9-4397-83a5-865cf0dd293f-kube-api-access-99gr2\") pod \"ovn-controller-metrics-zmqsz\" (UID: \"b5204157-80d9-4397-83a5-865cf0dd293f\") " pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.859909 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-96lzm"] Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.861959 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.864299 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-zmqsz" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.864840 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.875105 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c323eb01-f638-4ec8-98a4-711fbd9063f2-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.875319 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c323eb01-f638-4ec8-98a4-711fbd9063f2-scripts\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.875505 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c323eb01-f638-4ec8-98a4-711fbd9063f2-config\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.875573 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtld5\" (UniqueName: \"kubernetes.io/projected/c323eb01-f638-4ec8-98a4-711fbd9063f2-kube-api-access-qtld5\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.875686 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c323eb01-f638-4ec8-98a4-711fbd9063f2-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.875866 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c323eb01-f638-4ec8-98a4-711fbd9063f2-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.875959 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c323eb01-f638-4ec8-98a4-711fbd9063f2-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.876373 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c323eb01-f638-4ec8-98a4-711fbd9063f2-scripts\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.877071 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c323eb01-f638-4ec8-98a4-711fbd9063f2-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.880769 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c323eb01-f638-4ec8-98a4-711fbd9063f2-config\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.883437 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c323eb01-f638-4ec8-98a4-711fbd9063f2-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.884455 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c323eb01-f638-4ec8-98a4-711fbd9063f2-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.889891 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-96lzm"] Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.891778 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c323eb01-f638-4ec8-98a4-711fbd9063f2-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.906051 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtld5\" (UniqueName: \"kubernetes.io/projected/c323eb01-f638-4ec8-98a4-711fbd9063f2-kube-api-access-qtld5\") pod \"ovn-northd-0\" (UID: \"c323eb01-f638-4ec8-98a4-711fbd9063f2\") " pod="openstack/ovn-northd-0" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.978494 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-96lzm\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.978552 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-96lzm\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.978623 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-config\") pod \"dnsmasq-dns-8554648995-96lzm\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.978730 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8zzx\" (UniqueName: \"kubernetes.io/projected/a1d16fe0-89be-49c5-8282-6f7e20cdb417-kube-api-access-c8zzx\") pod \"dnsmasq-dns-8554648995-96lzm\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.979409 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-dns-svc\") pod \"dnsmasq-dns-8554648995-96lzm\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:30 crc kubenswrapper[4828]: I1210 19:18:30.990693 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.009953 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.082724 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8zzx\" (UniqueName: \"kubernetes.io/projected/a1d16fe0-89be-49c5-8282-6f7e20cdb417-kube-api-access-c8zzx\") pod \"dnsmasq-dns-8554648995-96lzm\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.083143 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-dns-svc\") pod \"dnsmasq-dns-8554648995-96lzm\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.083277 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-96lzm\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.083301 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-96lzm\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.083368 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-config\") pod \"dnsmasq-dns-8554648995-96lzm\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.085216 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-dns-svc\") pod \"dnsmasq-dns-8554648995-96lzm\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.085525 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-config\") pod \"dnsmasq-dns-8554648995-96lzm\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.085685 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-96lzm\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.096493 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-96lzm\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.116596 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ddf0d133-f122-4bac-b746-222a86e29dd2","Type":"ContainerStarted","Data":"5976d393bb7a64f50055c87c7a04004eb2c7947ff0021404c7ee2aacb9e52e73"} Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.140233 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8zzx\" (UniqueName: \"kubernetes.io/projected/a1d16fe0-89be-49c5-8282-6f7e20cdb417-kube-api-access-c8zzx\") pod \"dnsmasq-dns-8554648995-96lzm\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.154483 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=25.109512913 podStartE2EDuration="39.154461741s" podCreationTimestamp="2025-12-10 19:17:52 +0000 UTC" firstStartedPulling="2025-12-10 19:18:08.469714903 +0000 UTC m=+1368.980325908" lastFinishedPulling="2025-12-10 19:18:22.514663731 +0000 UTC m=+1383.025274736" observedRunningTime="2025-12-10 19:18:31.14666327 +0000 UTC m=+1391.657274265" watchObservedRunningTime="2025-12-10 19:18:31.154461741 +0000 UTC m=+1391.665072746" Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.227972 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.334417 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-8hxrs"] Dec 10 19:18:31 crc kubenswrapper[4828]: W1210 19:18:31.344781 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fa69bb6_5aa3_4f2e_b881_7445ea1922a2.slice/crio-0485329d4bb591c11ae6ee2cdfb9e457d7891d3a5e865142a1058d03c54a0f53 WatchSource:0}: Error finding container 0485329d4bb591c11ae6ee2cdfb9e457d7891d3a5e865142a1058d03c54a0f53: Status 404 returned error can't find the container with id 0485329d4bb591c11ae6ee2cdfb9e457d7891d3a5e865142a1058d03c54a0f53 Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.550633 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-zmqsz"] Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.658345 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 10 19:18:31 crc kubenswrapper[4828]: W1210 19:18:31.798672 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1d16fe0_89be_49c5_8282_6f7e20cdb417.slice/crio-3e78cdb4b0e5ad1220a2eaab680247f23e4180beb602052cd9d08fbfa48d1e87 WatchSource:0}: Error finding container 3e78cdb4b0e5ad1220a2eaab680247f23e4180beb602052cd9d08fbfa48d1e87: Status 404 returned error can't find the container with id 3e78cdb4b0e5ad1220a2eaab680247f23e4180beb602052cd9d08fbfa48d1e87 Dec 10 19:18:31 crc kubenswrapper[4828]: I1210 19:18:31.800818 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-96lzm"] Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.126395 4828 generic.go:334] "Generic (PLEG): container finished" podID="a1d16fe0-89be-49c5-8282-6f7e20cdb417" containerID="52fc67c13fce3212c06809daf8021cd32bb5ee841afa004575059931604350da" exitCode=0 Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.126462 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-96lzm" event={"ID":"a1d16fe0-89be-49c5-8282-6f7e20cdb417","Type":"ContainerDied","Data":"52fc67c13fce3212c06809daf8021cd32bb5ee841afa004575059931604350da"} Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.126495 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-96lzm" event={"ID":"a1d16fe0-89be-49c5-8282-6f7e20cdb417","Type":"ContainerStarted","Data":"3e78cdb4b0e5ad1220a2eaab680247f23e4180beb602052cd9d08fbfa48d1e87"} Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.129279 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-zmqsz" event={"ID":"b5204157-80d9-4397-83a5-865cf0dd293f","Type":"ContainerStarted","Data":"0125083bbcca666b421cb260616c165af0344366bd4998e3e5b7e18d32d43766"} Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.129356 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-zmqsz" event={"ID":"b5204157-80d9-4397-83a5-865cf0dd293f","Type":"ContainerStarted","Data":"2c974ec7fedb21ff334df9e9c4def96641f98ebebad2cc0933d48aeadcd556d5"} Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.130576 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c323eb01-f638-4ec8-98a4-711fbd9063f2","Type":"ContainerStarted","Data":"8f944a2c7b1bafc2e1aa9a3f850599b51044913576e426f4ae0b2ef430ff7418"} Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.134199 4828 generic.go:334] "Generic (PLEG): container finished" podID="1fa69bb6-5aa3-4f2e-b881-7445ea1922a2" containerID="21b5fc4e0d9e4dc704f8f3b6326a9d73501434ebe6fefd3c0668f2128d4ea9c5" exitCode=0 Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.134319 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" event={"ID":"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2","Type":"ContainerDied","Data":"21b5fc4e0d9e4dc704f8f3b6326a9d73501434ebe6fefd3c0668f2128d4ea9c5"} Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.134387 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" event={"ID":"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2","Type":"ContainerStarted","Data":"0485329d4bb591c11ae6ee2cdfb9e457d7891d3a5e865142a1058d03c54a0f53"} Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.256257 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-zmqsz" podStartSLOduration=2.256230961 podStartE2EDuration="2.256230961s" podCreationTimestamp="2025-12-10 19:18:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:32.179157589 +0000 UTC m=+1392.689768594" watchObservedRunningTime="2025-12-10 19:18:32.256230961 +0000 UTC m=+1392.766841966" Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.528388 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.621761 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-ovsdbserver-sb\") pod \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\" (UID: \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\") " Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.622076 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-config\") pod \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\" (UID: \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\") " Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.622152 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qshm\" (UniqueName: \"kubernetes.io/projected/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-kube-api-access-6qshm\") pod \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\" (UID: \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\") " Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.622198 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-dns-svc\") pod \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\" (UID: \"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2\") " Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.626926 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-kube-api-access-6qshm" (OuterVolumeSpecName: "kube-api-access-6qshm") pod "1fa69bb6-5aa3-4f2e-b881-7445ea1922a2" (UID: "1fa69bb6-5aa3-4f2e-b881-7445ea1922a2"). InnerVolumeSpecName "kube-api-access-6qshm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.645996 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1fa69bb6-5aa3-4f2e-b881-7445ea1922a2" (UID: "1fa69bb6-5aa3-4f2e-b881-7445ea1922a2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.648729 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-config" (OuterVolumeSpecName: "config") pod "1fa69bb6-5aa3-4f2e-b881-7445ea1922a2" (UID: "1fa69bb6-5aa3-4f2e-b881-7445ea1922a2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.659863 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1fa69bb6-5aa3-4f2e-b881-7445ea1922a2" (UID: "1fa69bb6-5aa3-4f2e-b881-7445ea1922a2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.724822 4828 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.725157 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.725173 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:32 crc kubenswrapper[4828]: I1210 19:18:32.725186 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qshm\" (UniqueName: \"kubernetes.io/projected/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2-kube-api-access-6qshm\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:33 crc kubenswrapper[4828]: I1210 19:18:33.152162 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" event={"ID":"1fa69bb6-5aa3-4f2e-b881-7445ea1922a2","Type":"ContainerDied","Data":"0485329d4bb591c11ae6ee2cdfb9e457d7891d3a5e865142a1058d03c54a0f53"} Dec 10 19:18:33 crc kubenswrapper[4828]: I1210 19:18:33.152224 4828 scope.go:117] "RemoveContainer" containerID="21b5fc4e0d9e4dc704f8f3b6326a9d73501434ebe6fefd3c0668f2128d4ea9c5" Dec 10 19:18:33 crc kubenswrapper[4828]: I1210 19:18:33.152367 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-8hxrs" Dec 10 19:18:33 crc kubenswrapper[4828]: I1210 19:18:33.156663 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-96lzm" event={"ID":"a1d16fe0-89be-49c5-8282-6f7e20cdb417","Type":"ContainerStarted","Data":"c08d10f7a026a12c2f25e55ed14b181cf80356c85fc43f4223f7af4d5215bdfe"} Dec 10 19:18:33 crc kubenswrapper[4828]: I1210 19:18:33.203597 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-96lzm" podStartSLOduration=3.203579738 podStartE2EDuration="3.203579738s" podCreationTimestamp="2025-12-10 19:18:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:33.182993351 +0000 UTC m=+1393.693604356" watchObservedRunningTime="2025-12-10 19:18:33.203579738 +0000 UTC m=+1393.714190743" Dec 10 19:18:33 crc kubenswrapper[4828]: I1210 19:18:33.273953 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-8hxrs"] Dec 10 19:18:33 crc kubenswrapper[4828]: I1210 19:18:33.289252 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-8hxrs"] Dec 10 19:18:33 crc kubenswrapper[4828]: I1210 19:18:33.799564 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fa69bb6-5aa3-4f2e-b881-7445ea1922a2" path="/var/lib/kubelet/pods/1fa69bb6-5aa3-4f2e-b881-7445ea1922a2/volumes" Dec 10 19:18:34 crc kubenswrapper[4828]: I1210 19:18:34.166947 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:34 crc kubenswrapper[4828]: I1210 19:18:34.222425 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 10 19:18:34 crc kubenswrapper[4828]: I1210 19:18:34.222477 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 10 19:18:35 crc kubenswrapper[4828]: I1210 19:18:35.658907 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 10 19:18:35 crc kubenswrapper[4828]: I1210 19:18:35.659385 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 10 19:18:36 crc kubenswrapper[4828]: I1210 19:18:36.184340 4828 generic.go:334] "Generic (PLEG): container finished" podID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerID="1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3" exitCode=0 Dec 10 19:18:36 crc kubenswrapper[4828]: I1210 19:18:36.184385 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0","Type":"ContainerDied","Data":"1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3"} Dec 10 19:18:37 crc kubenswrapper[4828]: I1210 19:18:37.194386 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c323eb01-f638-4ec8-98a4-711fbd9063f2","Type":"ContainerStarted","Data":"90c590b15fc89905ecd3ddb62b40635e3b3f564fff76868087ccf6b442868eed"} Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.274916 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.328865 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-96lzm"] Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.329408 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-96lzm" podUID="a1d16fe0-89be-49c5-8282-6f7e20cdb417" containerName="dnsmasq-dns" containerID="cri-o://c08d10f7a026a12c2f25e55ed14b181cf80356c85fc43f4223f7af4d5215bdfe" gracePeriod=10 Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.335289 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.415655 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-lh7pb"] Dec 10 19:18:38 crc kubenswrapper[4828]: E1210 19:18:38.416149 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fa69bb6-5aa3-4f2e-b881-7445ea1922a2" containerName="init" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.416168 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fa69bb6-5aa3-4f2e-b881-7445ea1922a2" containerName="init" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.416403 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fa69bb6-5aa3-4f2e-b881-7445ea1922a2" containerName="init" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.422893 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.443037 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-lh7pb"] Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.565159 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-lh7pb\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.565704 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-config\") pod \"dnsmasq-dns-b8fbc5445-lh7pb\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.565989 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-lh7pb\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.566044 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wkzn\" (UniqueName: \"kubernetes.io/projected/8e282513-5ba6-4eec-8738-8406c723a69b-kube-api-access-6wkzn\") pod \"dnsmasq-dns-b8fbc5445-lh7pb\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.566103 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-lh7pb\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.667593 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-lh7pb\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.667676 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-config\") pod \"dnsmasq-dns-b8fbc5445-lh7pb\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.667750 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-lh7pb\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.667781 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wkzn\" (UniqueName: \"kubernetes.io/projected/8e282513-5ba6-4eec-8738-8406c723a69b-kube-api-access-6wkzn\") pod \"dnsmasq-dns-b8fbc5445-lh7pb\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.667821 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-lh7pb\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.668736 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-lh7pb\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.668737 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-lh7pb\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.668851 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-config\") pod \"dnsmasq-dns-b8fbc5445-lh7pb\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.668990 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-lh7pb\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.696217 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wkzn\" (UniqueName: \"kubernetes.io/projected/8e282513-5ba6-4eec-8738-8406c723a69b-kube-api-access-6wkzn\") pod \"dnsmasq-dns-b8fbc5445-lh7pb\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.869853 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:38 crc kubenswrapper[4828]: I1210 19:18:38.969634 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.076463 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.223163 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c323eb01-f638-4ec8-98a4-711fbd9063f2","Type":"ContainerStarted","Data":"aeebaba8696fb42a2e1e3ad0b82e77a1443d9da8ddd1fc647d2495c37479242b"} Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.223291 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.239321 4828 generic.go:334] "Generic (PLEG): container finished" podID="a1d16fe0-89be-49c5-8282-6f7e20cdb417" containerID="c08d10f7a026a12c2f25e55ed14b181cf80356c85fc43f4223f7af4d5215bdfe" exitCode=0 Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.242939 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-96lzm" event={"ID":"a1d16fe0-89be-49c5-8282-6f7e20cdb417","Type":"ContainerDied","Data":"c08d10f7a026a12c2f25e55ed14b181cf80356c85fc43f4223f7af4d5215bdfe"} Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.248378 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=8.018521025 podStartE2EDuration="9.248359406s" podCreationTimestamp="2025-12-10 19:18:30 +0000 UTC" firstStartedPulling="2025-12-10 19:18:31.661733623 +0000 UTC m=+1392.172344628" lastFinishedPulling="2025-12-10 19:18:32.891572004 +0000 UTC m=+1393.402183009" observedRunningTime="2025-12-10 19:18:39.243183096 +0000 UTC m=+1399.753794101" watchObservedRunningTime="2025-12-10 19:18:39.248359406 +0000 UTC m=+1399.758970401" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.380266 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-lh7pb"] Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.440196 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.452638 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.456659 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-k7pr6" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.457905 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.458155 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.458453 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.465963 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.484905 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e29bd910-2077-4ebe-8028-e810061fb704-lock\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.485064 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.485093 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-788gr\" (UniqueName: \"kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-kube-api-access-788gr\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.485114 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.485136 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e29bd910-2077-4ebe-8028-e810061fb704-cache\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.586237 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.586282 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-788gr\" (UniqueName: \"kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-kube-api-access-788gr\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.586308 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.586330 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e29bd910-2077-4ebe-8028-e810061fb704-cache\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.586384 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e29bd910-2077-4ebe-8028-e810061fb704-lock\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.587017 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/swift-storage-0" Dec 10 19:18:39 crc kubenswrapper[4828]: E1210 19:18:39.589652 4828 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 19:18:39 crc kubenswrapper[4828]: E1210 19:18:39.589702 4828 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 19:18:39 crc kubenswrapper[4828]: E1210 19:18:39.589775 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift podName:e29bd910-2077-4ebe-8028-e810061fb704 nodeName:}" failed. No retries permitted until 2025-12-10 19:18:40.089751833 +0000 UTC m=+1400.600362918 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift") pod "swift-storage-0" (UID: "e29bd910-2077-4ebe-8028-e810061fb704") : configmap "swift-ring-files" not found Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.592029 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e29bd910-2077-4ebe-8028-e810061fb704-lock\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.594039 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e29bd910-2077-4ebe-8028-e810061fb704-cache\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.605200 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-788gr\" (UniqueName: \"kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-kube-api-access-788gr\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.614333 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.631253 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.790111 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-dns-svc\") pod \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.790476 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-ovsdbserver-nb\") pod \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.790543 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-config\") pod \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.790603 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-ovsdbserver-sb\") pod \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.790621 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8zzx\" (UniqueName: \"kubernetes.io/projected/a1d16fe0-89be-49c5-8282-6f7e20cdb417-kube-api-access-c8zzx\") pod \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\" (UID: \"a1d16fe0-89be-49c5-8282-6f7e20cdb417\") " Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.797551 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1d16fe0-89be-49c5-8282-6f7e20cdb417-kube-api-access-c8zzx" (OuterVolumeSpecName: "kube-api-access-c8zzx") pod "a1d16fe0-89be-49c5-8282-6f7e20cdb417" (UID: "a1d16fe0-89be-49c5-8282-6f7e20cdb417"). InnerVolumeSpecName "kube-api-access-c8zzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.836993 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a1d16fe0-89be-49c5-8282-6f7e20cdb417" (UID: "a1d16fe0-89be-49c5-8282-6f7e20cdb417"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.839405 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-config" (OuterVolumeSpecName: "config") pod "a1d16fe0-89be-49c5-8282-6f7e20cdb417" (UID: "a1d16fe0-89be-49c5-8282-6f7e20cdb417"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.843687 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a1d16fe0-89be-49c5-8282-6f7e20cdb417" (UID: "a1d16fe0-89be-49c5-8282-6f7e20cdb417"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.844317 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a1d16fe0-89be-49c5-8282-6f7e20cdb417" (UID: "a1d16fe0-89be-49c5-8282-6f7e20cdb417"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.894689 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.894748 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.894778 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8zzx\" (UniqueName: \"kubernetes.io/projected/a1d16fe0-89be-49c5-8282-6f7e20cdb417-kube-api-access-c8zzx\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.894789 4828 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.894823 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1d16fe0-89be-49c5-8282-6f7e20cdb417-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.940910 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-bw78w"] Dec 10 19:18:39 crc kubenswrapper[4828]: E1210 19:18:39.941400 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1d16fe0-89be-49c5-8282-6f7e20cdb417" containerName="init" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.941425 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1d16fe0-89be-49c5-8282-6f7e20cdb417" containerName="init" Dec 10 19:18:39 crc kubenswrapper[4828]: E1210 19:18:39.941452 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1d16fe0-89be-49c5-8282-6f7e20cdb417" containerName="dnsmasq-dns" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.941464 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1d16fe0-89be-49c5-8282-6f7e20cdb417" containerName="dnsmasq-dns" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.941699 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1d16fe0-89be-49c5-8282-6f7e20cdb417" containerName="dnsmasq-dns" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.942940 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.945177 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.945341 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.946615 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 10 19:18:39 crc kubenswrapper[4828]: I1210 19:18:39.953456 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bw78w"] Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.098573 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-swiftconf\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.098854 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vkzb\" (UniqueName: \"kubernetes.io/projected/66d31859-79eb-4314-9644-455cc33ddacd-kube-api-access-9vkzb\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.099015 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-combined-ca-bundle\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.099100 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.099178 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/66d31859-79eb-4314-9644-455cc33ddacd-etc-swift\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: E1210 19:18:40.099292 4828 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 19:18:40 crc kubenswrapper[4828]: E1210 19:18:40.099317 4828 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 19:18:40 crc kubenswrapper[4828]: E1210 19:18:40.099372 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift podName:e29bd910-2077-4ebe-8028-e810061fb704 nodeName:}" failed. No retries permitted until 2025-12-10 19:18:41.099354387 +0000 UTC m=+1401.609965392 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift") pod "swift-storage-0" (UID: "e29bd910-2077-4ebe-8028-e810061fb704") : configmap "swift-ring-files" not found Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.099457 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/66d31859-79eb-4314-9644-455cc33ddacd-ring-data-devices\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.099503 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66d31859-79eb-4314-9644-455cc33ddacd-scripts\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.100097 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-dispersionconf\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.201890 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vkzb\" (UniqueName: \"kubernetes.io/projected/66d31859-79eb-4314-9644-455cc33ddacd-kube-api-access-9vkzb\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.202007 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-combined-ca-bundle\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.202092 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/66d31859-79eb-4314-9644-455cc33ddacd-etc-swift\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.202216 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/66d31859-79eb-4314-9644-455cc33ddacd-ring-data-devices\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.202282 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66d31859-79eb-4314-9644-455cc33ddacd-scripts\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.202365 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-dispersionconf\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.202402 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-swiftconf\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.202851 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/66d31859-79eb-4314-9644-455cc33ddacd-etc-swift\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.203011 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/66d31859-79eb-4314-9644-455cc33ddacd-ring-data-devices\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.203226 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66d31859-79eb-4314-9644-455cc33ddacd-scripts\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.206119 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-swiftconf\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.206706 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-combined-ca-bundle\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.211832 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-dispersionconf\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.221231 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vkzb\" (UniqueName: \"kubernetes.io/projected/66d31859-79eb-4314-9644-455cc33ddacd-kube-api-access-9vkzb\") pod \"swift-ring-rebalance-bw78w\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.252635 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-96lzm" event={"ID":"a1d16fe0-89be-49c5-8282-6f7e20cdb417","Type":"ContainerDied","Data":"3e78cdb4b0e5ad1220a2eaab680247f23e4180beb602052cd9d08fbfa48d1e87"} Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.252644 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-96lzm" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.252690 4828 scope.go:117] "RemoveContainer" containerID="c08d10f7a026a12c2f25e55ed14b181cf80356c85fc43f4223f7af4d5215bdfe" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.254841 4828 generic.go:334] "Generic (PLEG): container finished" podID="8e282513-5ba6-4eec-8738-8406c723a69b" containerID="73fd9fac71d50343b5ae049985abafd0a4bbf2be45a1702f59cc2b82d9fd755a" exitCode=0 Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.254957 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" event={"ID":"8e282513-5ba6-4eec-8738-8406c723a69b","Type":"ContainerDied","Data":"73fd9fac71d50343b5ae049985abafd0a4bbf2be45a1702f59cc2b82d9fd755a"} Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.255006 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" event={"ID":"8e282513-5ba6-4eec-8738-8406c723a69b","Type":"ContainerStarted","Data":"9f5d1d75fd05a15d87c591fed65746e6786b2f811e1de06682e3a922e0eb0571"} Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.260617 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.320343 4828 scope.go:117] "RemoveContainer" containerID="52fc67c13fce3212c06809daf8021cd32bb5ee841afa004575059931604350da" Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.330947 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-96lzm"] Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.349602 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-96lzm"] Dec 10 19:18:40 crc kubenswrapper[4828]: I1210 19:18:40.795821 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bw78w"] Dec 10 19:18:41 crc kubenswrapper[4828]: I1210 19:18:41.130239 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:41 crc kubenswrapper[4828]: E1210 19:18:41.130406 4828 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 19:18:41 crc kubenswrapper[4828]: E1210 19:18:41.130420 4828 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 19:18:41 crc kubenswrapper[4828]: E1210 19:18:41.130467 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift podName:e29bd910-2077-4ebe-8028-e810061fb704 nodeName:}" failed. No retries permitted until 2025-12-10 19:18:43.130454567 +0000 UTC m=+1403.641065572 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift") pod "swift-storage-0" (UID: "e29bd910-2077-4ebe-8028-e810061fb704") : configmap "swift-ring-files" not found Dec 10 19:18:41 crc kubenswrapper[4828]: I1210 19:18:41.276270 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" event={"ID":"8e282513-5ba6-4eec-8738-8406c723a69b","Type":"ContainerStarted","Data":"f4ec0d6ab2875bf85d211db7bb4df48fce108ac5720f693b4635f071c1bdb002"} Dec 10 19:18:41 crc kubenswrapper[4828]: I1210 19:18:41.276531 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:41 crc kubenswrapper[4828]: I1210 19:18:41.293921 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bw78w" event={"ID":"66d31859-79eb-4314-9644-455cc33ddacd","Type":"ContainerStarted","Data":"d55cd18b32b2b00c3639d48f10b63cf32cd950f1f06e0ffe5b781e96613ff39e"} Dec 10 19:18:41 crc kubenswrapper[4828]: I1210 19:18:41.305963 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" podStartSLOduration=3.305945201 podStartE2EDuration="3.305945201s" podCreationTimestamp="2025-12-10 19:18:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:41.298257853 +0000 UTC m=+1401.808868858" watchObservedRunningTime="2025-12-10 19:18:41.305945201 +0000 UTC m=+1401.816556206" Dec 10 19:18:41 crc kubenswrapper[4828]: I1210 19:18:41.818945 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1d16fe0-89be-49c5-8282-6f7e20cdb417" path="/var/lib/kubelet/pods/a1d16fe0-89be-49c5-8282-6f7e20cdb417/volumes" Dec 10 19:18:42 crc kubenswrapper[4828]: I1210 19:18:42.424721 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 10 19:18:42 crc kubenswrapper[4828]: I1210 19:18:42.516223 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 10 19:18:43 crc kubenswrapper[4828]: I1210 19:18:43.184726 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:43 crc kubenswrapper[4828]: E1210 19:18:43.185286 4828 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 19:18:43 crc kubenswrapper[4828]: E1210 19:18:43.185450 4828 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 19:18:43 crc kubenswrapper[4828]: E1210 19:18:43.185506 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift podName:e29bd910-2077-4ebe-8028-e810061fb704 nodeName:}" failed. No retries permitted until 2025-12-10 19:18:47.185487794 +0000 UTC m=+1407.696098799 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift") pod "swift-storage-0" (UID: "e29bd910-2077-4ebe-8028-e810061fb704") : configmap "swift-ring-files" not found Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.117365 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-7994d84d8-5xz2j" podUID="5bd7de50-96a2-4ead-9ea2-f891902f61f1" containerName="console" containerID="cri-o://885ed7db92fb53103d12f67bbe8e1c5fa4606fb23015c598be39762944f115fd" gracePeriod=15 Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.345784 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7994d84d8-5xz2j_5bd7de50-96a2-4ead-9ea2-f891902f61f1/console/0.log" Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.345866 4828 generic.go:334] "Generic (PLEG): container finished" podID="5bd7de50-96a2-4ead-9ea2-f891902f61f1" containerID="885ed7db92fb53103d12f67bbe8e1c5fa4606fb23015c598be39762944f115fd" exitCode=2 Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.345903 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7994d84d8-5xz2j" event={"ID":"5bd7de50-96a2-4ead-9ea2-f891902f61f1","Type":"ContainerDied","Data":"885ed7db92fb53103d12f67bbe8e1c5fa4606fb23015c598be39762944f115fd"} Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.641407 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8a56-account-create-update-9dg89"] Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.643071 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8a56-account-create-update-9dg89" Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.645267 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.649654 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8a56-account-create-update-9dg89"] Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.699192 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-6bchn"] Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.700968 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-6bchn" Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.709169 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-6bchn"] Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.739089 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmbnr\" (UniqueName: \"kubernetes.io/projected/66db2b53-0c7d-4165-b743-99bddafea231-kube-api-access-cmbnr\") pod \"keystone-8a56-account-create-update-9dg89\" (UID: \"66db2b53-0c7d-4165-b743-99bddafea231\") " pod="openstack/keystone-8a56-account-create-update-9dg89" Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.739160 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66db2b53-0c7d-4165-b743-99bddafea231-operator-scripts\") pod \"keystone-8a56-account-create-update-9dg89\" (UID: \"66db2b53-0c7d-4165-b743-99bddafea231\") " pod="openstack/keystone-8a56-account-create-update-9dg89" Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.841506 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a-operator-scripts\") pod \"keystone-db-create-6bchn\" (UID: \"73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a\") " pod="openstack/keystone-db-create-6bchn" Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.841740 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knfqh\" (UniqueName: \"kubernetes.io/projected/73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a-kube-api-access-knfqh\") pod \"keystone-db-create-6bchn\" (UID: \"73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a\") " pod="openstack/keystone-db-create-6bchn" Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.841892 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmbnr\" (UniqueName: \"kubernetes.io/projected/66db2b53-0c7d-4165-b743-99bddafea231-kube-api-access-cmbnr\") pod \"keystone-8a56-account-create-update-9dg89\" (UID: \"66db2b53-0c7d-4165-b743-99bddafea231\") " pod="openstack/keystone-8a56-account-create-update-9dg89" Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.841973 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66db2b53-0c7d-4165-b743-99bddafea231-operator-scripts\") pod \"keystone-8a56-account-create-update-9dg89\" (UID: \"66db2b53-0c7d-4165-b743-99bddafea231\") " pod="openstack/keystone-8a56-account-create-update-9dg89" Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.843053 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66db2b53-0c7d-4165-b743-99bddafea231-operator-scripts\") pod \"keystone-8a56-account-create-update-9dg89\" (UID: \"66db2b53-0c7d-4165-b743-99bddafea231\") " pod="openstack/keystone-8a56-account-create-update-9dg89" Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.865149 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmbnr\" (UniqueName: \"kubernetes.io/projected/66db2b53-0c7d-4165-b743-99bddafea231-kube-api-access-cmbnr\") pod \"keystone-8a56-account-create-update-9dg89\" (UID: \"66db2b53-0c7d-4165-b743-99bddafea231\") " pod="openstack/keystone-8a56-account-create-update-9dg89" Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.944474 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a-operator-scripts\") pod \"keystone-db-create-6bchn\" (UID: \"73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a\") " pod="openstack/keystone-db-create-6bchn" Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.944642 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knfqh\" (UniqueName: \"kubernetes.io/projected/73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a-kube-api-access-knfqh\") pod \"keystone-db-create-6bchn\" (UID: \"73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a\") " pod="openstack/keystone-db-create-6bchn" Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.967634 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8a56-account-create-update-9dg89" Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.974166 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knfqh\" (UniqueName: \"kubernetes.io/projected/73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a-kube-api-access-knfqh\") pod \"keystone-db-create-6bchn\" (UID: \"73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a\") " pod="openstack/keystone-db-create-6bchn" Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.981118 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a-operator-scripts\") pod \"keystone-db-create-6bchn\" (UID: \"73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a\") " pod="openstack/keystone-db-create-6bchn" Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.990476 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-z45cg"] Dec 10 19:18:45 crc kubenswrapper[4828]: I1210 19:18:45.992661 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-z45cg" Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.015405 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-z45cg"] Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.022084 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-6bchn" Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.101247 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-eb79-account-create-update-kx4qg"] Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.106071 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-eb79-account-create-update-kx4qg" Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.108152 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.126468 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-eb79-account-create-update-kx4qg"] Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.150903 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxzcl\" (UniqueName: \"kubernetes.io/projected/1218268b-0e3f-41d5-8e3e-cac3f9164f70-kube-api-access-rxzcl\") pod \"placement-db-create-z45cg\" (UID: \"1218268b-0e3f-41d5-8e3e-cac3f9164f70\") " pod="openstack/placement-db-create-z45cg" Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.151106 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1218268b-0e3f-41d5-8e3e-cac3f9164f70-operator-scripts\") pod \"placement-db-create-z45cg\" (UID: \"1218268b-0e3f-41d5-8e3e-cac3f9164f70\") " pod="openstack/placement-db-create-z45cg" Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.252869 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qms7\" (UniqueName: \"kubernetes.io/projected/3fc92de3-2c9e-4e10-8d89-a71ffd49523d-kube-api-access-5qms7\") pod \"placement-eb79-account-create-update-kx4qg\" (UID: \"3fc92de3-2c9e-4e10-8d89-a71ffd49523d\") " pod="openstack/placement-eb79-account-create-update-kx4qg" Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.252924 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fc92de3-2c9e-4e10-8d89-a71ffd49523d-operator-scripts\") pod \"placement-eb79-account-create-update-kx4qg\" (UID: \"3fc92de3-2c9e-4e10-8d89-a71ffd49523d\") " pod="openstack/placement-eb79-account-create-update-kx4qg" Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.253062 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxzcl\" (UniqueName: \"kubernetes.io/projected/1218268b-0e3f-41d5-8e3e-cac3f9164f70-kube-api-access-rxzcl\") pod \"placement-db-create-z45cg\" (UID: \"1218268b-0e3f-41d5-8e3e-cac3f9164f70\") " pod="openstack/placement-db-create-z45cg" Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.253188 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1218268b-0e3f-41d5-8e3e-cac3f9164f70-operator-scripts\") pod \"placement-db-create-z45cg\" (UID: \"1218268b-0e3f-41d5-8e3e-cac3f9164f70\") " pod="openstack/placement-db-create-z45cg" Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.253903 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1218268b-0e3f-41d5-8e3e-cac3f9164f70-operator-scripts\") pod \"placement-db-create-z45cg\" (UID: \"1218268b-0e3f-41d5-8e3e-cac3f9164f70\") " pod="openstack/placement-db-create-z45cg" Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.270581 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxzcl\" (UniqueName: \"kubernetes.io/projected/1218268b-0e3f-41d5-8e3e-cac3f9164f70-kube-api-access-rxzcl\") pod \"placement-db-create-z45cg\" (UID: \"1218268b-0e3f-41d5-8e3e-cac3f9164f70\") " pod="openstack/placement-db-create-z45cg" Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.340248 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-z45cg" Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.355526 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qms7\" (UniqueName: \"kubernetes.io/projected/3fc92de3-2c9e-4e10-8d89-a71ffd49523d-kube-api-access-5qms7\") pod \"placement-eb79-account-create-update-kx4qg\" (UID: \"3fc92de3-2c9e-4e10-8d89-a71ffd49523d\") " pod="openstack/placement-eb79-account-create-update-kx4qg" Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.355582 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fc92de3-2c9e-4e10-8d89-a71ffd49523d-operator-scripts\") pod \"placement-eb79-account-create-update-kx4qg\" (UID: \"3fc92de3-2c9e-4e10-8d89-a71ffd49523d\") " pod="openstack/placement-eb79-account-create-update-kx4qg" Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.356391 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fc92de3-2c9e-4e10-8d89-a71ffd49523d-operator-scripts\") pod \"placement-eb79-account-create-update-kx4qg\" (UID: \"3fc92de3-2c9e-4e10-8d89-a71ffd49523d\") " pod="openstack/placement-eb79-account-create-update-kx4qg" Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.374678 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qms7\" (UniqueName: \"kubernetes.io/projected/3fc92de3-2c9e-4e10-8d89-a71ffd49523d-kube-api-access-5qms7\") pod \"placement-eb79-account-create-update-kx4qg\" (UID: \"3fc92de3-2c9e-4e10-8d89-a71ffd49523d\") " pod="openstack/placement-eb79-account-create-update-kx4qg" Dec 10 19:18:46 crc kubenswrapper[4828]: I1210 19:18:46.427427 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-eb79-account-create-update-kx4qg" Dec 10 19:18:47 crc kubenswrapper[4828]: I1210 19:18:47.274843 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:47 crc kubenswrapper[4828]: E1210 19:18:47.275061 4828 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 19:18:47 crc kubenswrapper[4828]: E1210 19:18:47.275238 4828 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 19:18:47 crc kubenswrapper[4828]: E1210 19:18:47.275300 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift podName:e29bd910-2077-4ebe-8028-e810061fb704 nodeName:}" failed. No retries permitted until 2025-12-10 19:18:55.275281869 +0000 UTC m=+1415.785892874 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift") pod "swift-storage-0" (UID: "e29bd910-2077-4ebe-8028-e810061fb704") : configmap "swift-ring-files" not found Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.073097 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-vd4h8"] Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.074312 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-vd4h8" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.112931 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-vd4h8"] Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.192456 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78s79\" (UniqueName: \"kubernetes.io/projected/32ccef7c-5acd-4ae1-a555-8079eaa96ae9-kube-api-access-78s79\") pod \"mysqld-exporter-openstack-db-create-vd4h8\" (UID: \"32ccef7c-5acd-4ae1-a555-8079eaa96ae9\") " pod="openstack/mysqld-exporter-openstack-db-create-vd4h8" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.192543 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32ccef7c-5acd-4ae1-a555-8079eaa96ae9-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-vd4h8\" (UID: \"32ccef7c-5acd-4ae1-a555-8079eaa96ae9\") " pod="openstack/mysqld-exporter-openstack-db-create-vd4h8" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.296308 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78s79\" (UniqueName: \"kubernetes.io/projected/32ccef7c-5acd-4ae1-a555-8079eaa96ae9-kube-api-access-78s79\") pod \"mysqld-exporter-openstack-db-create-vd4h8\" (UID: \"32ccef7c-5acd-4ae1-a555-8079eaa96ae9\") " pod="openstack/mysqld-exporter-openstack-db-create-vd4h8" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.297449 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32ccef7c-5acd-4ae1-a555-8079eaa96ae9-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-vd4h8\" (UID: \"32ccef7c-5acd-4ae1-a555-8079eaa96ae9\") " pod="openstack/mysqld-exporter-openstack-db-create-vd4h8" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.298654 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32ccef7c-5acd-4ae1-a555-8079eaa96ae9-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-vd4h8\" (UID: \"32ccef7c-5acd-4ae1-a555-8079eaa96ae9\") " pod="openstack/mysqld-exporter-openstack-db-create-vd4h8" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.315697 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0756-account-create-update-qr5d5"] Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.317638 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0756-account-create-update-qr5d5" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.319597 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-db-secret" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.323619 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78s79\" (UniqueName: \"kubernetes.io/projected/32ccef7c-5acd-4ae1-a555-8079eaa96ae9-kube-api-access-78s79\") pod \"mysqld-exporter-openstack-db-create-vd4h8\" (UID: \"32ccef7c-5acd-4ae1-a555-8079eaa96ae9\") " pod="openstack/mysqld-exporter-openstack-db-create-vd4h8" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.334962 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0756-account-create-update-qr5d5"] Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.372433 4828 generic.go:334] "Generic (PLEG): container finished" podID="fd6f7666-6cc8-478d-9acf-21cfbad541c9" containerID="9fb4907035796b871225aeb7750d239070e57fcc3199b135138f56a1c472ce23" exitCode=0 Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.372512 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fd6f7666-6cc8-478d-9acf-21cfbad541c9","Type":"ContainerDied","Data":"9fb4907035796b871225aeb7750d239070e57fcc3199b135138f56a1c472ce23"} Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.399571 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlcrh\" (UniqueName: \"kubernetes.io/projected/81aa6d69-1494-4790-a0b3-9e13444c17da-kube-api-access-rlcrh\") pod \"mysqld-exporter-0756-account-create-update-qr5d5\" (UID: \"81aa6d69-1494-4790-a0b3-9e13444c17da\") " pod="openstack/mysqld-exporter-0756-account-create-update-qr5d5" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.399861 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-vd4h8" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.400127 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81aa6d69-1494-4790-a0b3-9e13444c17da-operator-scripts\") pod \"mysqld-exporter-0756-account-create-update-qr5d5\" (UID: \"81aa6d69-1494-4790-a0b3-9e13444c17da\") " pod="openstack/mysqld-exporter-0756-account-create-update-qr5d5" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.502250 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlcrh\" (UniqueName: \"kubernetes.io/projected/81aa6d69-1494-4790-a0b3-9e13444c17da-kube-api-access-rlcrh\") pod \"mysqld-exporter-0756-account-create-update-qr5d5\" (UID: \"81aa6d69-1494-4790-a0b3-9e13444c17da\") " pod="openstack/mysqld-exporter-0756-account-create-update-qr5d5" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.502538 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81aa6d69-1494-4790-a0b3-9e13444c17da-operator-scripts\") pod \"mysqld-exporter-0756-account-create-update-qr5d5\" (UID: \"81aa6d69-1494-4790-a0b3-9e13444c17da\") " pod="openstack/mysqld-exporter-0756-account-create-update-qr5d5" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.503314 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81aa6d69-1494-4790-a0b3-9e13444c17da-operator-scripts\") pod \"mysqld-exporter-0756-account-create-update-qr5d5\" (UID: \"81aa6d69-1494-4790-a0b3-9e13444c17da\") " pod="openstack/mysqld-exporter-0756-account-create-update-qr5d5" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.518174 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlcrh\" (UniqueName: \"kubernetes.io/projected/81aa6d69-1494-4790-a0b3-9e13444c17da-kube-api-access-rlcrh\") pod \"mysqld-exporter-0756-account-create-update-qr5d5\" (UID: \"81aa6d69-1494-4790-a0b3-9e13444c17da\") " pod="openstack/mysqld-exporter-0756-account-create-update-qr5d5" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.757932 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0756-account-create-update-qr5d5" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.872995 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.947519 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4cpx8"] Dec 10 19:18:48 crc kubenswrapper[4828]: I1210 19:18:48.948545 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" podUID="d2630e12-86b3-45d3-84c0-dd50f5cf8c15" containerName="dnsmasq-dns" containerID="cri-o://a09f7f4c7627b83fbdb3841a2f8006809ec93cb5fda851ad0dbf804255d998eb" gracePeriod=10 Dec 10 19:18:49 crc kubenswrapper[4828]: I1210 19:18:49.383218 4828 generic.go:334] "Generic (PLEG): container finished" podID="b24f93de-d604-499b-be8e-2e934d119192" containerID="5505417dc4fa51ffbfa46576fe4ceb89ae8a05311568421b1979522bbc329385" exitCode=0 Dec 10 19:18:49 crc kubenswrapper[4828]: I1210 19:18:49.383287 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b24f93de-d604-499b-be8e-2e934d119192","Type":"ContainerDied","Data":"5505417dc4fa51ffbfa46576fe4ceb89ae8a05311568421b1979522bbc329385"} Dec 10 19:18:50 crc kubenswrapper[4828]: I1210 19:18:50.397838 4828 generic.go:334] "Generic (PLEG): container finished" podID="d2630e12-86b3-45d3-84c0-dd50f5cf8c15" containerID="a09f7f4c7627b83fbdb3841a2f8006809ec93cb5fda851ad0dbf804255d998eb" exitCode=0 Dec 10 19:18:50 crc kubenswrapper[4828]: I1210 19:18:50.398083 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" event={"ID":"d2630e12-86b3-45d3-84c0-dd50f5cf8c15","Type":"ContainerDied","Data":"a09f7f4c7627b83fbdb3841a2f8006809ec93cb5fda851ad0dbf804255d998eb"} Dec 10 19:18:50 crc kubenswrapper[4828]: I1210 19:18:50.951940 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7994d84d8-5xz2j_5bd7de50-96a2-4ead-9ea2-f891902f61f1/console/0.log" Dec 10 19:18:50 crc kubenswrapper[4828]: I1210 19:18:50.952446 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.058295 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-service-ca\") pod \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.058378 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-config\") pod \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.058459 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-oauth-serving-cert\") pod \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.058485 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-trusted-ca-bundle\") pod \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.058551 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-oauth-config\") pod \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.058634 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-serving-cert\") pod \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.058653 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72rch\" (UniqueName: \"kubernetes.io/projected/5bd7de50-96a2-4ead-9ea2-f891902f61f1-kube-api-access-72rch\") pod \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\" (UID: \"5bd7de50-96a2-4ead-9ea2-f891902f61f1\") " Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.059229 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-config" (OuterVolumeSpecName: "console-config") pod "5bd7de50-96a2-4ead-9ea2-f891902f61f1" (UID: "5bd7de50-96a2-4ead-9ea2-f891902f61f1"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.059240 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "5bd7de50-96a2-4ead-9ea2-f891902f61f1" (UID: "5bd7de50-96a2-4ead-9ea2-f891902f61f1"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.059253 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-service-ca" (OuterVolumeSpecName: "service-ca") pod "5bd7de50-96a2-4ead-9ea2-f891902f61f1" (UID: "5bd7de50-96a2-4ead-9ea2-f891902f61f1"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.059791 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "5bd7de50-96a2-4ead-9ea2-f891902f61f1" (UID: "5bd7de50-96a2-4ead-9ea2-f891902f61f1"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.059949 4828 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.059978 4828 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-service-ca\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.059989 4828 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.060001 4828 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5bd7de50-96a2-4ead-9ea2-f891902f61f1-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.063033 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "5bd7de50-96a2-4ead-9ea2-f891902f61f1" (UID: "5bd7de50-96a2-4ead-9ea2-f891902f61f1"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.063567 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.065479 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "5bd7de50-96a2-4ead-9ea2-f891902f61f1" (UID: "5bd7de50-96a2-4ead-9ea2-f891902f61f1"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.066777 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bd7de50-96a2-4ead-9ea2-f891902f61f1-kube-api-access-72rch" (OuterVolumeSpecName: "kube-api-access-72rch") pod "5bd7de50-96a2-4ead-9ea2-f891902f61f1" (UID: "5bd7de50-96a2-4ead-9ea2-f891902f61f1"). InnerVolumeSpecName "kube-api-access-72rch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.162066 4828 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.162127 4828 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5bd7de50-96a2-4ead-9ea2-f891902f61f1-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.162144 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72rch\" (UniqueName: \"kubernetes.io/projected/5bd7de50-96a2-4ead-9ea2-f891902f61f1-kube-api-access-72rch\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.220191 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.223415 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-74mkr"] Dec 10 19:18:51 crc kubenswrapper[4828]: E1210 19:18:51.223940 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2630e12-86b3-45d3-84c0-dd50f5cf8c15" containerName="dnsmasq-dns" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.224009 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2630e12-86b3-45d3-84c0-dd50f5cf8c15" containerName="dnsmasq-dns" Dec 10 19:18:51 crc kubenswrapper[4828]: E1210 19:18:51.224091 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bd7de50-96a2-4ead-9ea2-f891902f61f1" containerName="console" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.224141 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bd7de50-96a2-4ead-9ea2-f891902f61f1" containerName="console" Dec 10 19:18:51 crc kubenswrapper[4828]: E1210 19:18:51.224197 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2630e12-86b3-45d3-84c0-dd50f5cf8c15" containerName="init" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.224254 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2630e12-86b3-45d3-84c0-dd50f5cf8c15" containerName="init" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.224498 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bd7de50-96a2-4ead-9ea2-f891902f61f1" containerName="console" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.224586 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2630e12-86b3-45d3-84c0-dd50f5cf8c15" containerName="dnsmasq-dns" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.225601 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-74mkr" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.252120 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-74mkr"] Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.347976 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-bec3-account-create-update-gfhnk"] Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.356238 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-bec3-account-create-update-gfhnk" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.359471 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.366469 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-dns-svc\") pod \"d2630e12-86b3-45d3-84c0-dd50f5cf8c15\" (UID: \"d2630e12-86b3-45d3-84c0-dd50f5cf8c15\") " Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.366690 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sl49\" (UniqueName: \"kubernetes.io/projected/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-kube-api-access-5sl49\") pod \"d2630e12-86b3-45d3-84c0-dd50f5cf8c15\" (UID: \"d2630e12-86b3-45d3-84c0-dd50f5cf8c15\") " Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.366742 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-config\") pod \"d2630e12-86b3-45d3-84c0-dd50f5cf8c15\" (UID: \"d2630e12-86b3-45d3-84c0-dd50f5cf8c15\") " Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.367079 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e391ca51-f39a-4718-9265-4f21f26155e5-operator-scripts\") pod \"glance-db-create-74mkr\" (UID: \"e391ca51-f39a-4718-9265-4f21f26155e5\") " pod="openstack/glance-db-create-74mkr" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.367160 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbs4j\" (UniqueName: \"kubernetes.io/projected/e391ca51-f39a-4718-9265-4f21f26155e5-kube-api-access-xbs4j\") pod \"glance-db-create-74mkr\" (UID: \"e391ca51-f39a-4718-9265-4f21f26155e5\") " pod="openstack/glance-db-create-74mkr" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.373996 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-kube-api-access-5sl49" (OuterVolumeSpecName: "kube-api-access-5sl49") pod "d2630e12-86b3-45d3-84c0-dd50f5cf8c15" (UID: "d2630e12-86b3-45d3-84c0-dd50f5cf8c15"). InnerVolumeSpecName "kube-api-access-5sl49". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.397352 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-bec3-account-create-update-gfhnk"] Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.426094 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" event={"ID":"d2630e12-86b3-45d3-84c0-dd50f5cf8c15","Type":"ContainerDied","Data":"51e476d61488997d2c755a2a84bffd103a0f79f37735a8f8aa9875553841b706"} Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.426156 4828 scope.go:117] "RemoveContainer" containerID="a09f7f4c7627b83fbdb3841a2f8006809ec93cb5fda851ad0dbf804255d998eb" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.426302 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.435252 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7994d84d8-5xz2j_5bd7de50-96a2-4ead-9ea2-f891902f61f1/console/0.log" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.435307 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7994d84d8-5xz2j" event={"ID":"5bd7de50-96a2-4ead-9ea2-f891902f61f1","Type":"ContainerDied","Data":"a8925697c5ac2db98824d698d517c7a4ac5e9d33034b915c34fbfe7120225e35"} Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.435379 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7994d84d8-5xz2j" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.458568 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d2630e12-86b3-45d3-84c0-dd50f5cf8c15" (UID: "d2630e12-86b3-45d3-84c0-dd50f5cf8c15"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.469288 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/019b4ce6-57e1-42b0-be80-5c810a163d46-operator-scripts\") pod \"glance-bec3-account-create-update-gfhnk\" (UID: \"019b4ce6-57e1-42b0-be80-5c810a163d46\") " pod="openstack/glance-bec3-account-create-update-gfhnk" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.469389 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e391ca51-f39a-4718-9265-4f21f26155e5-operator-scripts\") pod \"glance-db-create-74mkr\" (UID: \"e391ca51-f39a-4718-9265-4f21f26155e5\") " pod="openstack/glance-db-create-74mkr" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.469465 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbs4j\" (UniqueName: \"kubernetes.io/projected/e391ca51-f39a-4718-9265-4f21f26155e5-kube-api-access-xbs4j\") pod \"glance-db-create-74mkr\" (UID: \"e391ca51-f39a-4718-9265-4f21f26155e5\") " pod="openstack/glance-db-create-74mkr" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.469493 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s8ns\" (UniqueName: \"kubernetes.io/projected/019b4ce6-57e1-42b0-be80-5c810a163d46-kube-api-access-4s8ns\") pod \"glance-bec3-account-create-update-gfhnk\" (UID: \"019b4ce6-57e1-42b0-be80-5c810a163d46\") " pod="openstack/glance-bec3-account-create-update-gfhnk" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.469568 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sl49\" (UniqueName: \"kubernetes.io/projected/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-kube-api-access-5sl49\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.469585 4828 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.471529 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e391ca51-f39a-4718-9265-4f21f26155e5-operator-scripts\") pod \"glance-db-create-74mkr\" (UID: \"e391ca51-f39a-4718-9265-4f21f26155e5\") " pod="openstack/glance-db-create-74mkr" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.494081 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-config" (OuterVolumeSpecName: "config") pod "d2630e12-86b3-45d3-84c0-dd50f5cf8c15" (UID: "d2630e12-86b3-45d3-84c0-dd50f5cf8c15"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.505771 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbs4j\" (UniqueName: \"kubernetes.io/projected/e391ca51-f39a-4718-9265-4f21f26155e5-kube-api-access-xbs4j\") pod \"glance-db-create-74mkr\" (UID: \"e391ca51-f39a-4718-9265-4f21f26155e5\") " pod="openstack/glance-db-create-74mkr" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.524931 4828 scope.go:117] "RemoveContainer" containerID="c4ad2ff657869f590f022f32220e87ccadf7b672abfe55db3e29bd7626616c1a" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.553132 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-74mkr" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.570776 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/019b4ce6-57e1-42b0-be80-5c810a163d46-operator-scripts\") pod \"glance-bec3-account-create-update-gfhnk\" (UID: \"019b4ce6-57e1-42b0-be80-5c810a163d46\") " pod="openstack/glance-bec3-account-create-update-gfhnk" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.570949 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s8ns\" (UniqueName: \"kubernetes.io/projected/019b4ce6-57e1-42b0-be80-5c810a163d46-kube-api-access-4s8ns\") pod \"glance-bec3-account-create-update-gfhnk\" (UID: \"019b4ce6-57e1-42b0-be80-5c810a163d46\") " pod="openstack/glance-bec3-account-create-update-gfhnk" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.571044 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d2630e12-86b3-45d3-84c0-dd50f5cf8c15-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.571403 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/019b4ce6-57e1-42b0-be80-5c810a163d46-operator-scripts\") pod \"glance-bec3-account-create-update-gfhnk\" (UID: \"019b4ce6-57e1-42b0-be80-5c810a163d46\") " pod="openstack/glance-bec3-account-create-update-gfhnk" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.579399 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-7994d84d8-5xz2j"] Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.581280 4828 scope.go:117] "RemoveContainer" containerID="885ed7db92fb53103d12f67bbe8e1c5fa4606fb23015c598be39762944f115fd" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.592237 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-7994d84d8-5xz2j"] Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.658629 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s8ns\" (UniqueName: \"kubernetes.io/projected/019b4ce6-57e1-42b0-be80-5c810a163d46-kube-api-access-4s8ns\") pod \"glance-bec3-account-create-update-gfhnk\" (UID: \"019b4ce6-57e1-42b0-be80-5c810a163d46\") " pod="openstack/glance-bec3-account-create-update-gfhnk" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.688273 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-bec3-account-create-update-gfhnk" Dec 10 19:18:51 crc kubenswrapper[4828]: I1210 19:18:51.831610 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bd7de50-96a2-4ead-9ea2-f891902f61f1" path="/var/lib/kubelet/pods/5bd7de50-96a2-4ead-9ea2-f891902f61f1/volumes" Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.372047 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-vd4h8"] Dec 10 19:18:52 crc kubenswrapper[4828]: W1210 19:18:52.375407 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32ccef7c_5acd_4ae1_a555_8079eaa96ae9.slice/crio-6cad7d9892a25e9813474be859f3265a1f3e9826c9a7f3766dd2af4ad3f081d2 WatchSource:0}: Error finding container 6cad7d9892a25e9813474be859f3265a1f3e9826c9a7f3766dd2af4ad3f081d2: Status 404 returned error can't find the container with id 6cad7d9892a25e9813474be859f3265a1f3e9826c9a7f3766dd2af4ad3f081d2 Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.387367 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8a56-account-create-update-9dg89"] Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.399600 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-z45cg"] Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.431422 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-6bchn"] Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.446843 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0756-account-create-update-qr5d5"] Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.448430 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8a56-account-create-update-9dg89" event={"ID":"66db2b53-0c7d-4165-b743-99bddafea231","Type":"ContainerStarted","Data":"c0753d13c8c966f11f9199828babd602b1ec4fb7030c9eeb600dcd10958e3608"} Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.453946 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0756-account-create-update-qr5d5" event={"ID":"81aa6d69-1494-4790-a0b3-9e13444c17da","Type":"ContainerStarted","Data":"9b117897435911d309b228475402730474a7a695ea1f7f49c56b14a058b35668"} Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.456707 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-vd4h8" event={"ID":"32ccef7c-5acd-4ae1-a555-8079eaa96ae9","Type":"ContainerStarted","Data":"6cad7d9892a25e9813474be859f3265a1f3e9826c9a7f3766dd2af4ad3f081d2"} Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.457124 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-eb79-account-create-update-kx4qg"] Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.457959 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-z45cg" event={"ID":"1218268b-0e3f-41d5-8e3e-cac3f9164f70","Type":"ContainerStarted","Data":"9b2d73001b2c9c12fc4cb4fe982bf9f9b72c358f3eea4f40e205d4c916b79442"} Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.461617 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0","Type":"ContainerStarted","Data":"93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b"} Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.468423 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-eb79-account-create-update-kx4qg" event={"ID":"3fc92de3-2c9e-4e10-8d89-a71ffd49523d","Type":"ContainerStarted","Data":"0ae97311368e2b21dd0f4dc9bfdc936694236cd2754e21302c1ed1ebcc9551e5"} Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.473761 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b24f93de-d604-499b-be8e-2e934d119192","Type":"ContainerStarted","Data":"cebed91d9e42c67e258a49b3b4f0266cf93e32df764a470725ba35c027b4f1af"} Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.474042 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.475998 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fd6f7666-6cc8-478d-9acf-21cfbad541c9","Type":"ContainerStarted","Data":"3cf3999e422c09e05f979b2209c544abc8e36d082b76293e4209e83d892753ba"} Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.476428 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.477146 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-6bchn" event={"ID":"73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a","Type":"ContainerStarted","Data":"d72ac7d4e8b31b1035ba9388c220a19f5691517eb72ef7ace30f11e59eeb6108"} Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.497070 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=40.896811317 podStartE2EDuration="1m1.497048591s" podCreationTimestamp="2025-12-10 19:17:51 +0000 UTC" firstStartedPulling="2025-12-10 19:17:53.314936347 +0000 UTC m=+1353.825547342" lastFinishedPulling="2025-12-10 19:18:13.915173611 +0000 UTC m=+1374.425784616" observedRunningTime="2025-12-10 19:18:52.493844725 +0000 UTC m=+1413.004455740" watchObservedRunningTime="2025-12-10 19:18:52.497048591 +0000 UTC m=+1413.007659596" Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.517741 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-74mkr"] Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.534456 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-bec3-account-create-update-gfhnk"] Dec 10 19:18:52 crc kubenswrapper[4828]: I1210 19:18:52.547981 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=41.709305078 podStartE2EDuration="1m2.5476924s" podCreationTimestamp="2025-12-10 19:17:50 +0000 UTC" firstStartedPulling="2025-12-10 19:17:53.146750181 +0000 UTC m=+1353.657361186" lastFinishedPulling="2025-12-10 19:18:13.985137503 +0000 UTC m=+1374.495748508" observedRunningTime="2025-12-10 19:18:52.522471469 +0000 UTC m=+1413.033082474" watchObservedRunningTime="2025-12-10 19:18:52.5476924 +0000 UTC m=+1413.058303405" Dec 10 19:18:52 crc kubenswrapper[4828]: W1210 19:18:52.593218 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode391ca51_f39a_4718_9265_4f21f26155e5.slice/crio-4e15dd59d76df32b9df4d813cea46b30a28346d66ec842311d565d7d0da8199c WatchSource:0}: Error finding container 4e15dd59d76df32b9df4d813cea46b30a28346d66ec842311d565d7d0da8199c: Status 404 returned error can't find the container with id 4e15dd59d76df32b9df4d813cea46b30a28346d66ec842311d565d7d0da8199c Dec 10 19:18:52 crc kubenswrapper[4828]: W1210 19:18:52.595167 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod019b4ce6_57e1_42b0_be80_5c810a163d46.slice/crio-367b8d383c2919522b8aba4024b6208f66721b597d42f023c7adadff0b54bccd WatchSource:0}: Error finding container 367b8d383c2919522b8aba4024b6208f66721b597d42f023c7adadff0b54bccd: Status 404 returned error can't find the container with id 367b8d383c2919522b8aba4024b6208f66721b597d42f023c7adadff0b54bccd Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.494812 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-eb79-account-create-update-kx4qg" event={"ID":"3fc92de3-2c9e-4e10-8d89-a71ffd49523d","Type":"ContainerStarted","Data":"c5d601775a8e90f10c6f5da090a2eac53127cbf90d0380f63d9f9454b0eccfee"} Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.501353 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-bec3-account-create-update-gfhnk" event={"ID":"019b4ce6-57e1-42b0-be80-5c810a163d46","Type":"ContainerStarted","Data":"88c45d03b97d34cdd7dab5b7b715fb5e59cf4330c07fa76ef60ff0c465365dc8"} Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.501398 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-bec3-account-create-update-gfhnk" event={"ID":"019b4ce6-57e1-42b0-be80-5c810a163d46","Type":"ContainerStarted","Data":"367b8d383c2919522b8aba4024b6208f66721b597d42f023c7adadff0b54bccd"} Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.505921 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-74mkr" event={"ID":"e391ca51-f39a-4718-9265-4f21f26155e5","Type":"ContainerStarted","Data":"ecd68268f1cafc889845b4718d6959187049f757ea6e70219c5c2ca365c91efc"} Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.505983 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-74mkr" event={"ID":"e391ca51-f39a-4718-9265-4f21f26155e5","Type":"ContainerStarted","Data":"4e15dd59d76df32b9df4d813cea46b30a28346d66ec842311d565d7d0da8199c"} Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.509569 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-6bchn" event={"ID":"73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a","Type":"ContainerStarted","Data":"0fea48dccbb4ec05cb44a530fb12900f0c322976b667d19f0ed7a0cba015e4a0"} Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.512066 4828 generic.go:334] "Generic (PLEG): container finished" podID="1218268b-0e3f-41d5-8e3e-cac3f9164f70" containerID="a0d7410b9dd4c96e017faa103058fdfd8d45257a1322ac9f2e0749cf50a1a584" exitCode=0 Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.512150 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-z45cg" event={"ID":"1218268b-0e3f-41d5-8e3e-cac3f9164f70","Type":"ContainerDied","Data":"a0d7410b9dd4c96e017faa103058fdfd8d45257a1322ac9f2e0749cf50a1a584"} Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.513741 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-vd4h8" event={"ID":"32ccef7c-5acd-4ae1-a555-8079eaa96ae9","Type":"ContainerStarted","Data":"b5590164ad17189bd48f55e46455646b366b4df0533a870dcaafbc1917e0cc96"} Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.516092 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bw78w" event={"ID":"66d31859-79eb-4314-9644-455cc33ddacd","Type":"ContainerStarted","Data":"d35f7a60aa187a3693db837f7c41fffad472b489f9e3f6e9f62738800da2a0a8"} Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.519458 4828 generic.go:334] "Generic (PLEG): container finished" podID="66db2b53-0c7d-4165-b743-99bddafea231" containerID="92c5c70edf42b6537b5effa2035053b811de539786aae6d531025d8b6ec7eb9c" exitCode=0 Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.519497 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8a56-account-create-update-9dg89" event={"ID":"66db2b53-0c7d-4165-b743-99bddafea231","Type":"ContainerDied","Data":"92c5c70edf42b6537b5effa2035053b811de539786aae6d531025d8b6ec7eb9c"} Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.522445 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0756-account-create-update-qr5d5" event={"ID":"81aa6d69-1494-4790-a0b3-9e13444c17da","Type":"ContainerStarted","Data":"4e3dd83d7be744622f33d7ee0d8f4daab774a0d444472c78b8b37f48caf17bde"} Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.527907 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-eb79-account-create-update-kx4qg" podStartSLOduration=7.527895345 podStartE2EDuration="7.527895345s" podCreationTimestamp="2025-12-10 19:18:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:53.525397807 +0000 UTC m=+1414.036008832" watchObservedRunningTime="2025-12-10 19:18:53.527895345 +0000 UTC m=+1414.038506350" Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.557355 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-bw78w" podStartSLOduration=4.088695209 podStartE2EDuration="14.557335531s" podCreationTimestamp="2025-12-10 19:18:39 +0000 UTC" firstStartedPulling="2025-12-10 19:18:40.802606596 +0000 UTC m=+1401.313217601" lastFinishedPulling="2025-12-10 19:18:51.271246918 +0000 UTC m=+1411.781857923" observedRunningTime="2025-12-10 19:18:53.555033928 +0000 UTC m=+1414.065644953" watchObservedRunningTime="2025-12-10 19:18:53.557335531 +0000 UTC m=+1414.067946536" Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.573424 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0756-account-create-update-qr5d5" podStartSLOduration=5.573404195 podStartE2EDuration="5.573404195s" podCreationTimestamp="2025-12-10 19:18:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:53.571455162 +0000 UTC m=+1414.082066157" watchObservedRunningTime="2025-12-10 19:18:53.573404195 +0000 UTC m=+1414.084015190" Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.600375 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-6bchn" podStartSLOduration=8.600358894 podStartE2EDuration="8.600358894s" podCreationTimestamp="2025-12-10 19:18:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:53.590573079 +0000 UTC m=+1414.101184084" watchObservedRunningTime="2025-12-10 19:18:53.600358894 +0000 UTC m=+1414.110969899" Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.617275 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-openstack-db-create-vd4h8" podStartSLOduration=5.61725548 podStartE2EDuration="5.61725548s" podCreationTimestamp="2025-12-10 19:18:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:53.612778089 +0000 UTC m=+1414.123389094" watchObservedRunningTime="2025-12-10 19:18:53.61725548 +0000 UTC m=+1414.127866485" Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.643378 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-74mkr" podStartSLOduration=2.643362756 podStartE2EDuration="2.643362756s" podCreationTimestamp="2025-12-10 19:18:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:53.636434349 +0000 UTC m=+1414.147045344" watchObservedRunningTime="2025-12-10 19:18:53.643362756 +0000 UTC m=+1414.153973761" Dec 10 19:18:53 crc kubenswrapper[4828]: I1210 19:18:53.657086 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-bec3-account-create-update-gfhnk" podStartSLOduration=2.657067537 podStartE2EDuration="2.657067537s" podCreationTimestamp="2025-12-10 19:18:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:53.655169175 +0000 UTC m=+1414.165780180" watchObservedRunningTime="2025-12-10 19:18:53.657067537 +0000 UTC m=+1414.167678542" Dec 10 19:18:54 crc kubenswrapper[4828]: I1210 19:18:54.531054 4828 generic.go:334] "Generic (PLEG): container finished" podID="3fc92de3-2c9e-4e10-8d89-a71ffd49523d" containerID="c5d601775a8e90f10c6f5da090a2eac53127cbf90d0380f63d9f9454b0eccfee" exitCode=0 Dec 10 19:18:54 crc kubenswrapper[4828]: I1210 19:18:54.531219 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-eb79-account-create-update-kx4qg" event={"ID":"3fc92de3-2c9e-4e10-8d89-a71ffd49523d","Type":"ContainerDied","Data":"c5d601775a8e90f10c6f5da090a2eac53127cbf90d0380f63d9f9454b0eccfee"} Dec 10 19:18:54 crc kubenswrapper[4828]: I1210 19:18:54.533100 4828 generic.go:334] "Generic (PLEG): container finished" podID="019b4ce6-57e1-42b0-be80-5c810a163d46" containerID="88c45d03b97d34cdd7dab5b7b715fb5e59cf4330c07fa76ef60ff0c465365dc8" exitCode=0 Dec 10 19:18:54 crc kubenswrapper[4828]: I1210 19:18:54.533162 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-bec3-account-create-update-gfhnk" event={"ID":"019b4ce6-57e1-42b0-be80-5c810a163d46","Type":"ContainerDied","Data":"88c45d03b97d34cdd7dab5b7b715fb5e59cf4330c07fa76ef60ff0c465365dc8"} Dec 10 19:18:54 crc kubenswrapper[4828]: I1210 19:18:54.534844 4828 generic.go:334] "Generic (PLEG): container finished" podID="81aa6d69-1494-4790-a0b3-9e13444c17da" containerID="4e3dd83d7be744622f33d7ee0d8f4daab774a0d444472c78b8b37f48caf17bde" exitCode=0 Dec 10 19:18:54 crc kubenswrapper[4828]: I1210 19:18:54.534909 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0756-account-create-update-qr5d5" event={"ID":"81aa6d69-1494-4790-a0b3-9e13444c17da","Type":"ContainerDied","Data":"4e3dd83d7be744622f33d7ee0d8f4daab774a0d444472c78b8b37f48caf17bde"} Dec 10 19:18:54 crc kubenswrapper[4828]: I1210 19:18:54.536544 4828 generic.go:334] "Generic (PLEG): container finished" podID="e391ca51-f39a-4718-9265-4f21f26155e5" containerID="ecd68268f1cafc889845b4718d6959187049f757ea6e70219c5c2ca365c91efc" exitCode=0 Dec 10 19:18:54 crc kubenswrapper[4828]: I1210 19:18:54.536595 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-74mkr" event={"ID":"e391ca51-f39a-4718-9265-4f21f26155e5","Type":"ContainerDied","Data":"ecd68268f1cafc889845b4718d6959187049f757ea6e70219c5c2ca365c91efc"} Dec 10 19:18:54 crc kubenswrapper[4828]: I1210 19:18:54.537702 4828 generic.go:334] "Generic (PLEG): container finished" podID="73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a" containerID="0fea48dccbb4ec05cb44a530fb12900f0c322976b667d19f0ed7a0cba015e4a0" exitCode=0 Dec 10 19:18:54 crc kubenswrapper[4828]: I1210 19:18:54.537742 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-6bchn" event={"ID":"73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a","Type":"ContainerDied","Data":"0fea48dccbb4ec05cb44a530fb12900f0c322976b667d19f0ed7a0cba015e4a0"} Dec 10 19:18:54 crc kubenswrapper[4828]: I1210 19:18:54.538834 4828 generic.go:334] "Generic (PLEG): container finished" podID="32ccef7c-5acd-4ae1-a555-8079eaa96ae9" containerID="b5590164ad17189bd48f55e46455646b366b4df0533a870dcaafbc1917e0cc96" exitCode=0 Dec 10 19:18:54 crc kubenswrapper[4828]: I1210 19:18:54.539771 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-vd4h8" event={"ID":"32ccef7c-5acd-4ae1-a555-8079eaa96ae9","Type":"ContainerDied","Data":"b5590164ad17189bd48f55e46455646b366b4df0533a870dcaafbc1917e0cc96"} Dec 10 19:18:54 crc kubenswrapper[4828]: I1210 19:18:54.952941 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-z45cg" Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.066718 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1218268b-0e3f-41d5-8e3e-cac3f9164f70-operator-scripts\") pod \"1218268b-0e3f-41d5-8e3e-cac3f9164f70\" (UID: \"1218268b-0e3f-41d5-8e3e-cac3f9164f70\") " Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.066886 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxzcl\" (UniqueName: \"kubernetes.io/projected/1218268b-0e3f-41d5-8e3e-cac3f9164f70-kube-api-access-rxzcl\") pod \"1218268b-0e3f-41d5-8e3e-cac3f9164f70\" (UID: \"1218268b-0e3f-41d5-8e3e-cac3f9164f70\") " Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.068238 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1218268b-0e3f-41d5-8e3e-cac3f9164f70-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1218268b-0e3f-41d5-8e3e-cac3f9164f70" (UID: "1218268b-0e3f-41d5-8e3e-cac3f9164f70"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.077103 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1218268b-0e3f-41d5-8e3e-cac3f9164f70-kube-api-access-rxzcl" (OuterVolumeSpecName: "kube-api-access-rxzcl") pod "1218268b-0e3f-41d5-8e3e-cac3f9164f70" (UID: "1218268b-0e3f-41d5-8e3e-cac3f9164f70"). InnerVolumeSpecName "kube-api-access-rxzcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.148383 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8a56-account-create-update-9dg89" Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.169493 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxzcl\" (UniqueName: \"kubernetes.io/projected/1218268b-0e3f-41d5-8e3e-cac3f9164f70-kube-api-access-rxzcl\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.169775 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1218268b-0e3f-41d5-8e3e-cac3f9164f70-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.270545 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66db2b53-0c7d-4165-b743-99bddafea231-operator-scripts\") pod \"66db2b53-0c7d-4165-b743-99bddafea231\" (UID: \"66db2b53-0c7d-4165-b743-99bddafea231\") " Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.270837 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmbnr\" (UniqueName: \"kubernetes.io/projected/66db2b53-0c7d-4165-b743-99bddafea231-kube-api-access-cmbnr\") pod \"66db2b53-0c7d-4165-b743-99bddafea231\" (UID: \"66db2b53-0c7d-4165-b743-99bddafea231\") " Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.270984 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66db2b53-0c7d-4165-b743-99bddafea231-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "66db2b53-0c7d-4165-b743-99bddafea231" (UID: "66db2b53-0c7d-4165-b743-99bddafea231"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.271342 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/66db2b53-0c7d-4165-b743-99bddafea231-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.283971 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66db2b53-0c7d-4165-b743-99bddafea231-kube-api-access-cmbnr" (OuterVolumeSpecName: "kube-api-access-cmbnr") pod "66db2b53-0c7d-4165-b743-99bddafea231" (UID: "66db2b53-0c7d-4165-b743-99bddafea231"). InnerVolumeSpecName "kube-api-access-cmbnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.373181 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.373382 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmbnr\" (UniqueName: \"kubernetes.io/projected/66db2b53-0c7d-4165-b743-99bddafea231-kube-api-access-cmbnr\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:55 crc kubenswrapper[4828]: E1210 19:18:55.373394 4828 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 10 19:18:55 crc kubenswrapper[4828]: E1210 19:18:55.373427 4828 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 10 19:18:55 crc kubenswrapper[4828]: E1210 19:18:55.373491 4828 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift podName:e29bd910-2077-4ebe-8028-e810061fb704 nodeName:}" failed. No retries permitted until 2025-12-10 19:19:11.37347264 +0000 UTC m=+1431.884083635 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift") pod "swift-storage-0" (UID: "e29bd910-2077-4ebe-8028-e810061fb704") : configmap "swift-ring-files" not found Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.548066 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-z45cg" Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.548059 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-z45cg" event={"ID":"1218268b-0e3f-41d5-8e3e-cac3f9164f70","Type":"ContainerDied","Data":"9b2d73001b2c9c12fc4cb4fe982bf9f9b72c358f3eea4f40e205d4c916b79442"} Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.548403 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b2d73001b2c9c12fc4cb4fe982bf9f9b72c358f3eea4f40e205d4c916b79442" Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.549552 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0","Type":"ContainerStarted","Data":"b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f"} Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.551007 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8a56-account-create-update-9dg89" event={"ID":"66db2b53-0c7d-4165-b743-99bddafea231","Type":"ContainerDied","Data":"c0753d13c8c966f11f9199828babd602b1ec4fb7030c9eeb600dcd10958e3608"} Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.551046 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0753d13c8c966f11f9199828babd602b1ec4fb7030c9eeb600dcd10958e3608" Dec 10 19:18:55 crc kubenswrapper[4828]: I1210 19:18:55.551107 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8a56-account-create-update-9dg89" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.078397 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-vd4h8" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.186125 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-74mkr" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.191923 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32ccef7c-5acd-4ae1-a555-8079eaa96ae9-operator-scripts\") pod \"32ccef7c-5acd-4ae1-a555-8079eaa96ae9\" (UID: \"32ccef7c-5acd-4ae1-a555-8079eaa96ae9\") " Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.192068 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78s79\" (UniqueName: \"kubernetes.io/projected/32ccef7c-5acd-4ae1-a555-8079eaa96ae9-kube-api-access-78s79\") pod \"32ccef7c-5acd-4ae1-a555-8079eaa96ae9\" (UID: \"32ccef7c-5acd-4ae1-a555-8079eaa96ae9\") " Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.208031 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32ccef7c-5acd-4ae1-a555-8079eaa96ae9-kube-api-access-78s79" (OuterVolumeSpecName: "kube-api-access-78s79") pod "32ccef7c-5acd-4ae1-a555-8079eaa96ae9" (UID: "32ccef7c-5acd-4ae1-a555-8079eaa96ae9"). InnerVolumeSpecName "kube-api-access-78s79". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.209395 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32ccef7c-5acd-4ae1-a555-8079eaa96ae9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "32ccef7c-5acd-4ae1-a555-8079eaa96ae9" (UID: "32ccef7c-5acd-4ae1-a555-8079eaa96ae9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.257375 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-eb79-account-create-update-kx4qg" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.274544 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-bec3-account-create-update-gfhnk" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.279692 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0756-account-create-update-qr5d5" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.293847 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fc92de3-2c9e-4e10-8d89-a71ffd49523d-operator-scripts\") pod \"3fc92de3-2c9e-4e10-8d89-a71ffd49523d\" (UID: \"3fc92de3-2c9e-4e10-8d89-a71ffd49523d\") " Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.293955 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e391ca51-f39a-4718-9265-4f21f26155e5-operator-scripts\") pod \"e391ca51-f39a-4718-9265-4f21f26155e5\" (UID: \"e391ca51-f39a-4718-9265-4f21f26155e5\") " Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.294058 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbs4j\" (UniqueName: \"kubernetes.io/projected/e391ca51-f39a-4718-9265-4f21f26155e5-kube-api-access-xbs4j\") pod \"e391ca51-f39a-4718-9265-4f21f26155e5\" (UID: \"e391ca51-f39a-4718-9265-4f21f26155e5\") " Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.294100 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qms7\" (UniqueName: \"kubernetes.io/projected/3fc92de3-2c9e-4e10-8d89-a71ffd49523d-kube-api-access-5qms7\") pod \"3fc92de3-2c9e-4e10-8d89-a71ffd49523d\" (UID: \"3fc92de3-2c9e-4e10-8d89-a71ffd49523d\") " Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.294718 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/32ccef7c-5acd-4ae1-a555-8079eaa96ae9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.294752 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78s79\" (UniqueName: \"kubernetes.io/projected/32ccef7c-5acd-4ae1-a555-8079eaa96ae9-kube-api-access-78s79\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.295890 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e391ca51-f39a-4718-9265-4f21f26155e5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e391ca51-f39a-4718-9265-4f21f26155e5" (UID: "e391ca51-f39a-4718-9265-4f21f26155e5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.296316 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3fc92de3-2c9e-4e10-8d89-a71ffd49523d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3fc92de3-2c9e-4e10-8d89-a71ffd49523d" (UID: "3fc92de3-2c9e-4e10-8d89-a71ffd49523d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.297404 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-6bchn" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.299838 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e391ca51-f39a-4718-9265-4f21f26155e5-kube-api-access-xbs4j" (OuterVolumeSpecName: "kube-api-access-xbs4j") pod "e391ca51-f39a-4718-9265-4f21f26155e5" (UID: "e391ca51-f39a-4718-9265-4f21f26155e5"). InnerVolumeSpecName "kube-api-access-xbs4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.302204 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fc92de3-2c9e-4e10-8d89-a71ffd49523d-kube-api-access-5qms7" (OuterVolumeSpecName: "kube-api-access-5qms7") pod "3fc92de3-2c9e-4e10-8d89-a71ffd49523d" (UID: "3fc92de3-2c9e-4e10-8d89-a71ffd49523d"). InnerVolumeSpecName "kube-api-access-5qms7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.395560 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/019b4ce6-57e1-42b0-be80-5c810a163d46-operator-scripts\") pod \"019b4ce6-57e1-42b0-be80-5c810a163d46\" (UID: \"019b4ce6-57e1-42b0-be80-5c810a163d46\") " Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.395819 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81aa6d69-1494-4790-a0b3-9e13444c17da-operator-scripts\") pod \"81aa6d69-1494-4790-a0b3-9e13444c17da\" (UID: \"81aa6d69-1494-4790-a0b3-9e13444c17da\") " Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.395882 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlcrh\" (UniqueName: \"kubernetes.io/projected/81aa6d69-1494-4790-a0b3-9e13444c17da-kube-api-access-rlcrh\") pod \"81aa6d69-1494-4790-a0b3-9e13444c17da\" (UID: \"81aa6d69-1494-4790-a0b3-9e13444c17da\") " Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.395922 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a-operator-scripts\") pod \"73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a\" (UID: \"73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a\") " Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.395961 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4s8ns\" (UniqueName: \"kubernetes.io/projected/019b4ce6-57e1-42b0-be80-5c810a163d46-kube-api-access-4s8ns\") pod \"019b4ce6-57e1-42b0-be80-5c810a163d46\" (UID: \"019b4ce6-57e1-42b0-be80-5c810a163d46\") " Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.396024 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knfqh\" (UniqueName: \"kubernetes.io/projected/73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a-kube-api-access-knfqh\") pod \"73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a\" (UID: \"73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a\") " Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.396105 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/019b4ce6-57e1-42b0-be80-5c810a163d46-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "019b4ce6-57e1-42b0-be80-5c810a163d46" (UID: "019b4ce6-57e1-42b0-be80-5c810a163d46"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.396605 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbs4j\" (UniqueName: \"kubernetes.io/projected/e391ca51-f39a-4718-9265-4f21f26155e5-kube-api-access-xbs4j\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.396627 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qms7\" (UniqueName: \"kubernetes.io/projected/3fc92de3-2c9e-4e10-8d89-a71ffd49523d-kube-api-access-5qms7\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.396637 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/019b4ce6-57e1-42b0-be80-5c810a163d46-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.396647 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3fc92de3-2c9e-4e10-8d89-a71ffd49523d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.396658 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e391ca51-f39a-4718-9265-4f21f26155e5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.396778 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a" (UID: "73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.397386 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81aa6d69-1494-4790-a0b3-9e13444c17da-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "81aa6d69-1494-4790-a0b3-9e13444c17da" (UID: "81aa6d69-1494-4790-a0b3-9e13444c17da"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.400883 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a-kube-api-access-knfqh" (OuterVolumeSpecName: "kube-api-access-knfqh") pod "73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a" (UID: "73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a"). InnerVolumeSpecName "kube-api-access-knfqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.402073 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/019b4ce6-57e1-42b0-be80-5c810a163d46-kube-api-access-4s8ns" (OuterVolumeSpecName: "kube-api-access-4s8ns") pod "019b4ce6-57e1-42b0-be80-5c810a163d46" (UID: "019b4ce6-57e1-42b0-be80-5c810a163d46"). InnerVolumeSpecName "kube-api-access-4s8ns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.402958 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81aa6d69-1494-4790-a0b3-9e13444c17da-kube-api-access-rlcrh" (OuterVolumeSpecName: "kube-api-access-rlcrh") pod "81aa6d69-1494-4790-a0b3-9e13444c17da" (UID: "81aa6d69-1494-4790-a0b3-9e13444c17da"). InnerVolumeSpecName "kube-api-access-rlcrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.498817 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81aa6d69-1494-4790-a0b3-9e13444c17da-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.498848 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlcrh\" (UniqueName: \"kubernetes.io/projected/81aa6d69-1494-4790-a0b3-9e13444c17da-kube-api-access-rlcrh\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.498858 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.498867 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4s8ns\" (UniqueName: \"kubernetes.io/projected/019b4ce6-57e1-42b0-be80-5c810a163d46-kube-api-access-4s8ns\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.498877 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knfqh\" (UniqueName: \"kubernetes.io/projected/73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a-kube-api-access-knfqh\") on node \"crc\" DevicePath \"\"" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.561813 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-vd4h8" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.561859 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-vd4h8" event={"ID":"32ccef7c-5acd-4ae1-a555-8079eaa96ae9","Type":"ContainerDied","Data":"6cad7d9892a25e9813474be859f3265a1f3e9826c9a7f3766dd2af4ad3f081d2"} Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.561901 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cad7d9892a25e9813474be859f3265a1f3e9826c9a7f3766dd2af4ad3f081d2" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.563981 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-eb79-account-create-update-kx4qg" event={"ID":"3fc92de3-2c9e-4e10-8d89-a71ffd49523d","Type":"ContainerDied","Data":"0ae97311368e2b21dd0f4dc9bfdc936694236cd2754e21302c1ed1ebcc9551e5"} Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.564016 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ae97311368e2b21dd0f4dc9bfdc936694236cd2754e21302c1ed1ebcc9551e5" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.563996 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-eb79-account-create-update-kx4qg" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.569510 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-bec3-account-create-update-gfhnk" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.569630 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-bec3-account-create-update-gfhnk" event={"ID":"019b4ce6-57e1-42b0-be80-5c810a163d46","Type":"ContainerDied","Data":"367b8d383c2919522b8aba4024b6208f66721b597d42f023c7adadff0b54bccd"} Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.569690 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="367b8d383c2919522b8aba4024b6208f66721b597d42f023c7adadff0b54bccd" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.581639 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0756-account-create-update-qr5d5" event={"ID":"81aa6d69-1494-4790-a0b3-9e13444c17da","Type":"ContainerDied","Data":"9b117897435911d309b228475402730474a7a695ea1f7f49c56b14a058b35668"} Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.581663 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0756-account-create-update-qr5d5" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.581690 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b117897435911d309b228475402730474a7a695ea1f7f49c56b14a058b35668" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.586863 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-74mkr" event={"ID":"e391ca51-f39a-4718-9265-4f21f26155e5","Type":"ContainerDied","Data":"4e15dd59d76df32b9df4d813cea46b30a28346d66ec842311d565d7d0da8199c"} Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.586898 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e15dd59d76df32b9df4d813cea46b30a28346d66ec842311d565d7d0da8199c" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.586982 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-74mkr" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.592211 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-6bchn" event={"ID":"73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a","Type":"ContainerDied","Data":"d72ac7d4e8b31b1035ba9388c220a19f5691517eb72ef7ace30f11e59eeb6108"} Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.592256 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d72ac7d4e8b31b1035ba9388c220a19f5691517eb72ef7ace30f11e59eeb6108" Dec 10 19:18:56 crc kubenswrapper[4828]: I1210 19:18:56.592337 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-6bchn" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.134532 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-krphm" podUID="8fa35e66-ea0f-423d-b133-2170900e433f" containerName="ovn-controller" probeResult="failure" output=< Dec 10 19:18:57 crc kubenswrapper[4828]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 10 19:18:57 crc kubenswrapper[4828]: > Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.201863 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.210221 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-lldb4" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.450013 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-krphm-config-vjj54"] Dec 10 19:18:57 crc kubenswrapper[4828]: E1210 19:18:57.450398 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fc92de3-2c9e-4e10-8d89-a71ffd49523d" containerName="mariadb-account-create-update" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.450414 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fc92de3-2c9e-4e10-8d89-a71ffd49523d" containerName="mariadb-account-create-update" Dec 10 19:18:57 crc kubenswrapper[4828]: E1210 19:18:57.450428 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1218268b-0e3f-41d5-8e3e-cac3f9164f70" containerName="mariadb-database-create" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.450434 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1218268b-0e3f-41d5-8e3e-cac3f9164f70" containerName="mariadb-database-create" Dec 10 19:18:57 crc kubenswrapper[4828]: E1210 19:18:57.450449 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66db2b53-0c7d-4165-b743-99bddafea231" containerName="mariadb-account-create-update" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.450456 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="66db2b53-0c7d-4165-b743-99bddafea231" containerName="mariadb-account-create-update" Dec 10 19:18:57 crc kubenswrapper[4828]: E1210 19:18:57.450466 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a" containerName="mariadb-database-create" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.450473 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a" containerName="mariadb-database-create" Dec 10 19:18:57 crc kubenswrapper[4828]: E1210 19:18:57.450485 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e391ca51-f39a-4718-9265-4f21f26155e5" containerName="mariadb-database-create" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.450491 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="e391ca51-f39a-4718-9265-4f21f26155e5" containerName="mariadb-database-create" Dec 10 19:18:57 crc kubenswrapper[4828]: E1210 19:18:57.450503 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019b4ce6-57e1-42b0-be80-5c810a163d46" containerName="mariadb-account-create-update" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.450509 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="019b4ce6-57e1-42b0-be80-5c810a163d46" containerName="mariadb-account-create-update" Dec 10 19:18:57 crc kubenswrapper[4828]: E1210 19:18:57.450524 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32ccef7c-5acd-4ae1-a555-8079eaa96ae9" containerName="mariadb-database-create" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.450530 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="32ccef7c-5acd-4ae1-a555-8079eaa96ae9" containerName="mariadb-database-create" Dec 10 19:18:57 crc kubenswrapper[4828]: E1210 19:18:57.450539 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81aa6d69-1494-4790-a0b3-9e13444c17da" containerName="mariadb-account-create-update" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.450545 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="81aa6d69-1494-4790-a0b3-9e13444c17da" containerName="mariadb-account-create-update" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.450736 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a" containerName="mariadb-database-create" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.450751 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="e391ca51-f39a-4718-9265-4f21f26155e5" containerName="mariadb-database-create" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.450763 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fc92de3-2c9e-4e10-8d89-a71ffd49523d" containerName="mariadb-account-create-update" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.450773 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1218268b-0e3f-41d5-8e3e-cac3f9164f70" containerName="mariadb-database-create" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.450782 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="81aa6d69-1494-4790-a0b3-9e13444c17da" containerName="mariadb-account-create-update" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.450791 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="32ccef7c-5acd-4ae1-a555-8079eaa96ae9" containerName="mariadb-database-create" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.450821 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="019b4ce6-57e1-42b0-be80-5c810a163d46" containerName="mariadb-account-create-update" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.450835 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="66db2b53-0c7d-4165-b743-99bddafea231" containerName="mariadb-account-create-update" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.451591 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.454248 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.464851 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-krphm-config-vjj54"] Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.528212 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84bz2\" (UniqueName: \"kubernetes.io/projected/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-kube-api-access-84bz2\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.528326 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-run\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.528422 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-additional-scripts\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.528602 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-log-ovn\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.528627 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-scripts\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.528712 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-run-ovn\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.630354 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84bz2\" (UniqueName: \"kubernetes.io/projected/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-kube-api-access-84bz2\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.630489 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-run\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.630514 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-additional-scripts\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.630582 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-log-ovn\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.630604 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-scripts\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.630643 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-run-ovn\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.630823 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-run\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.630876 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-run-ovn\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.630930 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-log-ovn\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.631714 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-additional-scripts\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.632588 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-scripts\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.658334 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84bz2\" (UniqueName: \"kubernetes.io/projected/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-kube-api-access-84bz2\") pod \"ovn-controller-krphm-config-vjj54\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:57 crc kubenswrapper[4828]: I1210 19:18:57.768975 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.594164 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m"] Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.595710 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m" Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.635888 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m"] Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.656535 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zchx5\" (UniqueName: \"kubernetes.io/projected/fb2e9f6f-100c-494a-bbca-44c3ab8fc46c-kube-api-access-zchx5\") pod \"mysqld-exporter-openstack-cell1-db-create-6nb4m\" (UID: \"fb2e9f6f-100c-494a-bbca-44c3ab8fc46c\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m" Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.656653 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb2e9f6f-100c-494a-bbca-44c3ab8fc46c-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-6nb4m\" (UID: \"fb2e9f6f-100c-494a-bbca-44c3ab8fc46c\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m" Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.758760 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zchx5\" (UniqueName: \"kubernetes.io/projected/fb2e9f6f-100c-494a-bbca-44c3ab8fc46c-kube-api-access-zchx5\") pod \"mysqld-exporter-openstack-cell1-db-create-6nb4m\" (UID: \"fb2e9f6f-100c-494a-bbca-44c3ab8fc46c\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m" Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.759156 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb2e9f6f-100c-494a-bbca-44c3ab8fc46c-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-6nb4m\" (UID: \"fb2e9f6f-100c-494a-bbca-44c3ab8fc46c\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m" Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.760072 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb2e9f6f-100c-494a-bbca-44c3ab8fc46c-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-6nb4m\" (UID: \"fb2e9f6f-100c-494a-bbca-44c3ab8fc46c\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m" Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.780514 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zchx5\" (UniqueName: \"kubernetes.io/projected/fb2e9f6f-100c-494a-bbca-44c3ab8fc46c-kube-api-access-zchx5\") pod \"mysqld-exporter-openstack-cell1-db-create-6nb4m\" (UID: \"fb2e9f6f-100c-494a-bbca-44c3ab8fc46c\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m" Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.796829 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-7f0a-account-create-update-nh472"] Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.798662 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-7f0a-account-create-update-nh472" Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.806663 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-cell1-db-secret" Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.808852 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-7f0a-account-create-update-nh472"] Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.860998 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr2bg\" (UniqueName: \"kubernetes.io/projected/ecab69d1-3989-4e69-90f3-b54d4e229283-kube-api-access-tr2bg\") pod \"mysqld-exporter-7f0a-account-create-update-nh472\" (UID: \"ecab69d1-3989-4e69-90f3-b54d4e229283\") " pod="openstack/mysqld-exporter-7f0a-account-create-update-nh472" Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.861059 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecab69d1-3989-4e69-90f3-b54d4e229283-operator-scripts\") pod \"mysqld-exporter-7f0a-account-create-update-nh472\" (UID: \"ecab69d1-3989-4e69-90f3-b54d4e229283\") " pod="openstack/mysqld-exporter-7f0a-account-create-update-nh472" Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.883071 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-krphm-config-vjj54"] Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.922500 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m" Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.963176 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr2bg\" (UniqueName: \"kubernetes.io/projected/ecab69d1-3989-4e69-90f3-b54d4e229283-kube-api-access-tr2bg\") pod \"mysqld-exporter-7f0a-account-create-update-nh472\" (UID: \"ecab69d1-3989-4e69-90f3-b54d4e229283\") " pod="openstack/mysqld-exporter-7f0a-account-create-update-nh472" Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.963233 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecab69d1-3989-4e69-90f3-b54d4e229283-operator-scripts\") pod \"mysqld-exporter-7f0a-account-create-update-nh472\" (UID: \"ecab69d1-3989-4e69-90f3-b54d4e229283\") " pod="openstack/mysqld-exporter-7f0a-account-create-update-nh472" Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.964179 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecab69d1-3989-4e69-90f3-b54d4e229283-operator-scripts\") pod \"mysqld-exporter-7f0a-account-create-update-nh472\" (UID: \"ecab69d1-3989-4e69-90f3-b54d4e229283\") " pod="openstack/mysqld-exporter-7f0a-account-create-update-nh472" Dec 10 19:18:58 crc kubenswrapper[4828]: I1210 19:18:58.984717 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr2bg\" (UniqueName: \"kubernetes.io/projected/ecab69d1-3989-4e69-90f3-b54d4e229283-kube-api-access-tr2bg\") pod \"mysqld-exporter-7f0a-account-create-update-nh472\" (UID: \"ecab69d1-3989-4e69-90f3-b54d4e229283\") " pod="openstack/mysqld-exporter-7f0a-account-create-update-nh472" Dec 10 19:18:59 crc kubenswrapper[4828]: I1210 19:18:59.151979 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-7f0a-account-create-update-nh472" Dec 10 19:18:59 crc kubenswrapper[4828]: I1210 19:18:59.548051 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m"] Dec 10 19:18:59 crc kubenswrapper[4828]: I1210 19:18:59.621350 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-krphm-config-vjj54" event={"ID":"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871","Type":"ContainerStarted","Data":"5a03b2ce640c72ef01f72c38f50700150bd2f332dd20556ad43d0e5d5291d8ad"} Dec 10 19:18:59 crc kubenswrapper[4828]: I1210 19:18:59.621411 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-krphm-config-vjj54" event={"ID":"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871","Type":"ContainerStarted","Data":"90af37df2fc65a27e735c06fb9b031a7fb218e0c885f72007478b823e707fa6d"} Dec 10 19:18:59 crc kubenswrapper[4828]: I1210 19:18:59.624539 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0","Type":"ContainerStarted","Data":"20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4"} Dec 10 19:18:59 crc kubenswrapper[4828]: I1210 19:18:59.626091 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m" event={"ID":"fb2e9f6f-100c-494a-bbca-44c3ab8fc46c","Type":"ContainerStarted","Data":"066485973a0982cf75622f246feaed1aa700cf9b26093a6c9505192526a3e6c1"} Dec 10 19:18:59 crc kubenswrapper[4828]: I1210 19:18:59.641201 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-krphm-config-vjj54" podStartSLOduration=2.641182934 podStartE2EDuration="2.641182934s" podCreationTimestamp="2025-12-10 19:18:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:18:59.636542499 +0000 UTC m=+1420.147153514" watchObservedRunningTime="2025-12-10 19:18:59.641182934 +0000 UTC m=+1420.151793939" Dec 10 19:18:59 crc kubenswrapper[4828]: I1210 19:18:59.662852 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=18.164501189 podStartE2EDuration="1m1.662832899s" podCreationTimestamp="2025-12-10 19:17:58 +0000 UTC" firstStartedPulling="2025-12-10 19:18:14.915116719 +0000 UTC m=+1375.425727724" lastFinishedPulling="2025-12-10 19:18:58.413448429 +0000 UTC m=+1418.924059434" observedRunningTime="2025-12-10 19:18:59.661092962 +0000 UTC m=+1420.171703977" watchObservedRunningTime="2025-12-10 19:18:59.662832899 +0000 UTC m=+1420.173443904" Dec 10 19:18:59 crc kubenswrapper[4828]: I1210 19:18:59.696733 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 10 19:18:59 crc kubenswrapper[4828]: I1210 19:18:59.696775 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 10 19:18:59 crc kubenswrapper[4828]: I1210 19:18:59.698855 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 10 19:18:59 crc kubenswrapper[4828]: I1210 19:18:59.732970 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-7f0a-account-create-update-nh472"] Dec 10 19:18:59 crc kubenswrapper[4828]: W1210 19:18:59.737137 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecab69d1_3989_4e69_90f3_b54d4e229283.slice/crio-0ca6be7b8faba33fad47e29dcea825f20b8995bf6094401b156fee7f1f46c55e WatchSource:0}: Error finding container 0ca6be7b8faba33fad47e29dcea825f20b8995bf6094401b156fee7f1f46c55e: Status 404 returned error can't find the container with id 0ca6be7b8faba33fad47e29dcea825f20b8995bf6094401b156fee7f1f46c55e Dec 10 19:19:00 crc kubenswrapper[4828]: I1210 19:19:00.635780 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-7f0a-account-create-update-nh472" event={"ID":"ecab69d1-3989-4e69-90f3-b54d4e229283","Type":"ContainerStarted","Data":"0ca6be7b8faba33fad47e29dcea825f20b8995bf6094401b156fee7f1f46c55e"} Dec 10 19:19:00 crc kubenswrapper[4828]: I1210 19:19:00.638031 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.647593 4828 generic.go:334] "Generic (PLEG): container finished" podID="1a97f1c5-2d06-4e5c-8b51-8d0a77a89871" containerID="5a03b2ce640c72ef01f72c38f50700150bd2f332dd20556ad43d0e5d5291d8ad" exitCode=0 Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.647682 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-krphm-config-vjj54" event={"ID":"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871","Type":"ContainerDied","Data":"5a03b2ce640c72ef01f72c38f50700150bd2f332dd20556ad43d0e5d5291d8ad"} Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.652522 4828 generic.go:334] "Generic (PLEG): container finished" podID="fb2e9f6f-100c-494a-bbca-44c3ab8fc46c" containerID="6d3ae3c549af589ec6acd20f78c025c3bd7121f7dfd42814870195fbffa23d55" exitCode=0 Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.652610 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m" event={"ID":"fb2e9f6f-100c-494a-bbca-44c3ab8fc46c","Type":"ContainerDied","Data":"6d3ae3c549af589ec6acd20f78c025c3bd7121f7dfd42814870195fbffa23d55"} Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.655205 4828 generic.go:334] "Generic (PLEG): container finished" podID="ecab69d1-3989-4e69-90f3-b54d4e229283" containerID="2a7168e8a3b73a3f90371d4052dccae3569a9f926e27b7fd2e18b2a8a57b63f4" exitCode=0 Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.655281 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-7f0a-account-create-update-nh472" event={"ID":"ecab69d1-3989-4e69-90f3-b54d4e229283","Type":"ContainerDied","Data":"2a7168e8a3b73a3f90371d4052dccae3569a9f926e27b7fd2e18b2a8a57b63f4"} Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.670480 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-9rgqj"] Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.671947 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9rgqj" Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.674710 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.674713 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-mdn2k" Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.681619 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-9rgqj"] Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.858772 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-config-data\") pod \"glance-db-sync-9rgqj\" (UID: \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\") " pod="openstack/glance-db-sync-9rgqj" Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.858945 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-combined-ca-bundle\") pod \"glance-db-sync-9rgqj\" (UID: \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\") " pod="openstack/glance-db-sync-9rgqj" Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.860222 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-db-sync-config-data\") pod \"glance-db-sync-9rgqj\" (UID: \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\") " pod="openstack/glance-db-sync-9rgqj" Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.860426 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bm6g\" (UniqueName: \"kubernetes.io/projected/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-kube-api-access-6bm6g\") pod \"glance-db-sync-9rgqj\" (UID: \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\") " pod="openstack/glance-db-sync-9rgqj" Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.963910 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-db-sync-config-data\") pod \"glance-db-sync-9rgqj\" (UID: \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\") " pod="openstack/glance-db-sync-9rgqj" Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.964028 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bm6g\" (UniqueName: \"kubernetes.io/projected/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-kube-api-access-6bm6g\") pod \"glance-db-sync-9rgqj\" (UID: \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\") " pod="openstack/glance-db-sync-9rgqj" Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.964086 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-config-data\") pod \"glance-db-sync-9rgqj\" (UID: \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\") " pod="openstack/glance-db-sync-9rgqj" Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.964178 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-combined-ca-bundle\") pod \"glance-db-sync-9rgqj\" (UID: \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\") " pod="openstack/glance-db-sync-9rgqj" Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.970553 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-combined-ca-bundle\") pod \"glance-db-sync-9rgqj\" (UID: \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\") " pod="openstack/glance-db-sync-9rgqj" Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.972331 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-config-data\") pod \"glance-db-sync-9rgqj\" (UID: \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\") " pod="openstack/glance-db-sync-9rgqj" Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.973111 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-db-sync-config-data\") pod \"glance-db-sync-9rgqj\" (UID: \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\") " pod="openstack/glance-db-sync-9rgqj" Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.987501 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bm6g\" (UniqueName: \"kubernetes.io/projected/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-kube-api-access-6bm6g\") pod \"glance-db-sync-9rgqj\" (UID: \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\") " pod="openstack/glance-db-sync-9rgqj" Dec 10 19:19:01 crc kubenswrapper[4828]: I1210 19:19:01.993213 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9rgqj" Dec 10 19:19:02 crc kubenswrapper[4828]: I1210 19:19:02.145573 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-krphm" Dec 10 19:19:02 crc kubenswrapper[4828]: I1210 19:19:02.572555 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="fd6f7666-6cc8-478d-9acf-21cfbad541c9" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.129:5671: connect: connection refused" Dec 10 19:19:02 crc kubenswrapper[4828]: I1210 19:19:02.585559 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-9rgqj"] Dec 10 19:19:02 crc kubenswrapper[4828]: I1210 19:19:02.655967 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="b24f93de-d604-499b-be8e-2e934d119192" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.130:5671: connect: connection refused" Dec 10 19:19:02 crc kubenswrapper[4828]: I1210 19:19:02.667501 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9rgqj" event={"ID":"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2","Type":"ContainerStarted","Data":"a728cb42c785d2ddde237828275982bd43f39677e51e04e1f2a777c387be51c1"} Dec 10 19:19:02 crc kubenswrapper[4828]: I1210 19:19:02.669636 4828 generic.go:334] "Generic (PLEG): container finished" podID="66d31859-79eb-4314-9644-455cc33ddacd" containerID="d35f7a60aa187a3693db837f7c41fffad472b489f9e3f6e9f62738800da2a0a8" exitCode=0 Dec 10 19:19:02 crc kubenswrapper[4828]: I1210 19:19:02.669806 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bw78w" event={"ID":"66d31859-79eb-4314-9644-455cc33ddacd","Type":"ContainerDied","Data":"d35f7a60aa187a3693db837f7c41fffad472b489f9e3f6e9f62738800da2a0a8"} Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.320601 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.403518 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb2e9f6f-100c-494a-bbca-44c3ab8fc46c-operator-scripts\") pod \"fb2e9f6f-100c-494a-bbca-44c3ab8fc46c\" (UID: \"fb2e9f6f-100c-494a-bbca-44c3ab8fc46c\") " Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.403939 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zchx5\" (UniqueName: \"kubernetes.io/projected/fb2e9f6f-100c-494a-bbca-44c3ab8fc46c-kube-api-access-zchx5\") pod \"fb2e9f6f-100c-494a-bbca-44c3ab8fc46c\" (UID: \"fb2e9f6f-100c-494a-bbca-44c3ab8fc46c\") " Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.404572 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb2e9f6f-100c-494a-bbca-44c3ab8fc46c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fb2e9f6f-100c-494a-bbca-44c3ab8fc46c" (UID: "fb2e9f6f-100c-494a-bbca-44c3ab8fc46c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.405169 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb2e9f6f-100c-494a-bbca-44c3ab8fc46c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.412241 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb2e9f6f-100c-494a-bbca-44c3ab8fc46c-kube-api-access-zchx5" (OuterVolumeSpecName: "kube-api-access-zchx5") pod "fb2e9f6f-100c-494a-bbca-44c3ab8fc46c" (UID: "fb2e9f6f-100c-494a-bbca-44c3ab8fc46c"). InnerVolumeSpecName "kube-api-access-zchx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.503920 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.506676 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zchx5\" (UniqueName: \"kubernetes.io/projected/fb2e9f6f-100c-494a-bbca-44c3ab8fc46c-kube-api-access-zchx5\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.511668 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-7f0a-account-create-update-nh472" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.609617 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-run\") pod \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.609685 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84bz2\" (UniqueName: \"kubernetes.io/projected/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-kube-api-access-84bz2\") pod \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.609723 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecab69d1-3989-4e69-90f3-b54d4e229283-operator-scripts\") pod \"ecab69d1-3989-4e69-90f3-b54d4e229283\" (UID: \"ecab69d1-3989-4e69-90f3-b54d4e229283\") " Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.609907 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tr2bg\" (UniqueName: \"kubernetes.io/projected/ecab69d1-3989-4e69-90f3-b54d4e229283-kube-api-access-tr2bg\") pod \"ecab69d1-3989-4e69-90f3-b54d4e229283\" (UID: \"ecab69d1-3989-4e69-90f3-b54d4e229283\") " Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.609927 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-log-ovn\") pod \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.610003 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-scripts\") pod \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.610040 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-run-ovn\") pod \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.610119 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-run" (OuterVolumeSpecName: "var-run") pod "1a97f1c5-2d06-4e5c-8b51-8d0a77a89871" (UID: "1a97f1c5-2d06-4e5c-8b51-8d0a77a89871"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.610261 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "1a97f1c5-2d06-4e5c-8b51-8d0a77a89871" (UID: "1a97f1c5-2d06-4e5c-8b51-8d0a77a89871"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.610523 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-additional-scripts\") pod \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\" (UID: \"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871\") " Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.611109 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecab69d1-3989-4e69-90f3-b54d4e229283-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ecab69d1-3989-4e69-90f3-b54d4e229283" (UID: "ecab69d1-3989-4e69-90f3-b54d4e229283"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.612598 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "1a97f1c5-2d06-4e5c-8b51-8d0a77a89871" (UID: "1a97f1c5-2d06-4e5c-8b51-8d0a77a89871"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.613300 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "1a97f1c5-2d06-4e5c-8b51-8d0a77a89871" (UID: "1a97f1c5-2d06-4e5c-8b51-8d0a77a89871"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.614705 4828 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-run\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.614981 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-scripts" (OuterVolumeSpecName: "scripts") pod "1a97f1c5-2d06-4e5c-8b51-8d0a77a89871" (UID: "1a97f1c5-2d06-4e5c-8b51-8d0a77a89871"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.615559 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecab69d1-3989-4e69-90f3-b54d4e229283-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.615584 4828 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.615602 4828 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.615611 4828 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.617328 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-kube-api-access-84bz2" (OuterVolumeSpecName: "kube-api-access-84bz2") pod "1a97f1c5-2d06-4e5c-8b51-8d0a77a89871" (UID: "1a97f1c5-2d06-4e5c-8b51-8d0a77a89871"). InnerVolumeSpecName "kube-api-access-84bz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.618153 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecab69d1-3989-4e69-90f3-b54d4e229283-kube-api-access-tr2bg" (OuterVolumeSpecName: "kube-api-access-tr2bg") pod "ecab69d1-3989-4e69-90f3-b54d4e229283" (UID: "ecab69d1-3989-4e69-90f3-b54d4e229283"). InnerVolumeSpecName "kube-api-access-tr2bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.637940 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.698534 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m" event={"ID":"fb2e9f6f-100c-494a-bbca-44c3ab8fc46c","Type":"ContainerDied","Data":"066485973a0982cf75622f246feaed1aa700cf9b26093a6c9505192526a3e6c1"} Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.698582 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="066485973a0982cf75622f246feaed1aa700cf9b26093a6c9505192526a3e6c1" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.698654 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.733109 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-7f0a-account-create-update-nh472" event={"ID":"ecab69d1-3989-4e69-90f3-b54d4e229283","Type":"ContainerDied","Data":"0ca6be7b8faba33fad47e29dcea825f20b8995bf6094401b156fee7f1f46c55e"} Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.733375 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ca6be7b8faba33fad47e29dcea825f20b8995bf6094401b156fee7f1f46c55e" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.733503 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-7f0a-account-create-update-nh472" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.734305 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tr2bg\" (UniqueName: \"kubernetes.io/projected/ecab69d1-3989-4e69-90f3-b54d4e229283-kube-api-access-tr2bg\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.735064 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.735085 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84bz2\" (UniqueName: \"kubernetes.io/projected/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871-kube-api-access-84bz2\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.737562 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-krphm-config-vjj54" event={"ID":"1a97f1c5-2d06-4e5c-8b51-8d0a77a89871","Type":"ContainerDied","Data":"90af37df2fc65a27e735c06fb9b031a7fb218e0c885f72007478b823e707fa6d"} Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.737594 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90af37df2fc65a27e735c06fb9b031a7fb218e0c885f72007478b823e707fa6d" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.737834 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-krphm-config-vjj54" Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.738895 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerName="prometheus" containerID="cri-o://93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b" gracePeriod=600 Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.739039 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerName="thanos-sidecar" containerID="cri-o://20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4" gracePeriod=600 Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.739091 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerName="config-reloader" containerID="cri-o://b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f" gracePeriod=600 Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.852353 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-krphm-config-vjj54"] Dec 10 19:19:03 crc kubenswrapper[4828]: I1210 19:19:03.895121 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-krphm-config-vjj54"] Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.279630 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.380878 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-dispersionconf\") pod \"66d31859-79eb-4314-9644-455cc33ddacd\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.381032 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/66d31859-79eb-4314-9644-455cc33ddacd-ring-data-devices\") pod \"66d31859-79eb-4314-9644-455cc33ddacd\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.381717 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66d31859-79eb-4314-9644-455cc33ddacd-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "66d31859-79eb-4314-9644-455cc33ddacd" (UID: "66d31859-79eb-4314-9644-455cc33ddacd"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.381834 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-combined-ca-bundle\") pod \"66d31859-79eb-4314-9644-455cc33ddacd\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.382347 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/66d31859-79eb-4314-9644-455cc33ddacd-etc-swift\") pod \"66d31859-79eb-4314-9644-455cc33ddacd\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.385313 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66d31859-79eb-4314-9644-455cc33ddacd-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "66d31859-79eb-4314-9644-455cc33ddacd" (UID: "66d31859-79eb-4314-9644-455cc33ddacd"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.385583 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66d31859-79eb-4314-9644-455cc33ddacd-scripts\") pod \"66d31859-79eb-4314-9644-455cc33ddacd\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.385626 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vkzb\" (UniqueName: \"kubernetes.io/projected/66d31859-79eb-4314-9644-455cc33ddacd-kube-api-access-9vkzb\") pod \"66d31859-79eb-4314-9644-455cc33ddacd\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.385672 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-swiftconf\") pod \"66d31859-79eb-4314-9644-455cc33ddacd\" (UID: \"66d31859-79eb-4314-9644-455cc33ddacd\") " Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.389597 4828 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/66d31859-79eb-4314-9644-455cc33ddacd-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.389624 4828 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/66d31859-79eb-4314-9644-455cc33ddacd-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.389788 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66d31859-79eb-4314-9644-455cc33ddacd-kube-api-access-9vkzb" (OuterVolumeSpecName: "kube-api-access-9vkzb") pod "66d31859-79eb-4314-9644-455cc33ddacd" (UID: "66d31859-79eb-4314-9644-455cc33ddacd"). InnerVolumeSpecName "kube-api-access-9vkzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.391155 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "66d31859-79eb-4314-9644-455cc33ddacd" (UID: "66d31859-79eb-4314-9644-455cc33ddacd"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.412704 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66d31859-79eb-4314-9644-455cc33ddacd-scripts" (OuterVolumeSpecName: "scripts") pod "66d31859-79eb-4314-9644-455cc33ddacd" (UID: "66d31859-79eb-4314-9644-455cc33ddacd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.429163 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "66d31859-79eb-4314-9644-455cc33ddacd" (UID: "66d31859-79eb-4314-9644-455cc33ddacd"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.451172 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66d31859-79eb-4314-9644-455cc33ddacd" (UID: "66d31859-79eb-4314-9644-455cc33ddacd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.461709 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.492170 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.492202 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66d31859-79eb-4314-9644-455cc33ddacd-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.492212 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vkzb\" (UniqueName: \"kubernetes.io/projected/66d31859-79eb-4314-9644-455cc33ddacd-kube-api-access-9vkzb\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.492222 4828 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.492231 4828 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/66d31859-79eb-4314-9644-455cc33ddacd-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.595084 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-prometheus-metric-storage-rulefiles-0\") pod \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.595233 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-config\") pod \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.595280 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-web-config\") pod \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.595365 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dppqj\" (UniqueName: \"kubernetes.io/projected/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-kube-api-access-dppqj\") pod \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.595390 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-config-out\") pod \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.595437 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-tls-assets\") pod \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.595456 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-thanos-prometheus-http-client-file\") pod \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.595478 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\" (UID: \"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0\") " Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.596130 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" (UID: "975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.596725 4828 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.598103 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-config" (OuterVolumeSpecName: "config") pod "975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" (UID: "975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.599860 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-config-out" (OuterVolumeSpecName: "config-out") pod "975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" (UID: "975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.603071 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-kube-api-access-dppqj" (OuterVolumeSpecName: "kube-api-access-dppqj") pod "975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" (UID: "975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0"). InnerVolumeSpecName "kube-api-access-dppqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.604154 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" (UID: "975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.604356 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" (UID: "975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.609321 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" (UID: "975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.627466 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-web-config" (OuterVolumeSpecName: "web-config") pod "975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" (UID: "975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.698499 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.698534 4828 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-web-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.698546 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dppqj\" (UniqueName: \"kubernetes.io/projected/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-kube-api-access-dppqj\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.698556 4828 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-config-out\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.698565 4828 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.698576 4828 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.698603 4828 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.721306 4828 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.750553 4828 generic.go:334] "Generic (PLEG): container finished" podID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerID="20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4" exitCode=0 Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.750588 4828 generic.go:334] "Generic (PLEG): container finished" podID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerID="b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f" exitCode=0 Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.750598 4828 generic.go:334] "Generic (PLEG): container finished" podID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerID="93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b" exitCode=0 Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.750643 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0","Type":"ContainerDied","Data":"20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4"} Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.750673 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0","Type":"ContainerDied","Data":"b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f"} Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.750687 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0","Type":"ContainerDied","Data":"93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b"} Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.750700 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0","Type":"ContainerDied","Data":"c8e245b3412b3b73e4489df2bdc8ee6671ab3b7fa58a6b1fd127953972800fed"} Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.750729 4828 scope.go:117] "RemoveContainer" containerID="20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.750925 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.754412 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bw78w" event={"ID":"66d31859-79eb-4314-9644-455cc33ddacd","Type":"ContainerDied","Data":"d55cd18b32b2b00c3639d48f10b63cf32cd950f1f06e0ffe5b781e96613ff39e"} Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.754484 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d55cd18b32b2b00c3639d48f10b63cf32cd950f1f06e0ffe5b781e96613ff39e" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.754552 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bw78w" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.778233 4828 scope.go:117] "RemoveContainer" containerID="b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.802154 4828 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.810313 4828 scope.go:117] "RemoveContainer" containerID="93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.817787 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.830482 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.845315 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:19:04 crc kubenswrapper[4828]: E1210 19:19:04.845893 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerName="init-config-reloader" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.845914 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerName="init-config-reloader" Dec 10 19:19:04 crc kubenswrapper[4828]: E1210 19:19:04.845933 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d31859-79eb-4314-9644-455cc33ddacd" containerName="swift-ring-rebalance" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.845941 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d31859-79eb-4314-9644-455cc33ddacd" containerName="swift-ring-rebalance" Dec 10 19:19:04 crc kubenswrapper[4828]: E1210 19:19:04.845954 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecab69d1-3989-4e69-90f3-b54d4e229283" containerName="mariadb-account-create-update" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.845962 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecab69d1-3989-4e69-90f3-b54d4e229283" containerName="mariadb-account-create-update" Dec 10 19:19:04 crc kubenswrapper[4828]: E1210 19:19:04.845986 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb2e9f6f-100c-494a-bbca-44c3ab8fc46c" containerName="mariadb-database-create" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.845993 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb2e9f6f-100c-494a-bbca-44c3ab8fc46c" containerName="mariadb-database-create" Dec 10 19:19:04 crc kubenswrapper[4828]: E1210 19:19:04.846012 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerName="config-reloader" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.846023 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerName="config-reloader" Dec 10 19:19:04 crc kubenswrapper[4828]: E1210 19:19:04.846035 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a97f1c5-2d06-4e5c-8b51-8d0a77a89871" containerName="ovn-config" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.846043 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a97f1c5-2d06-4e5c-8b51-8d0a77a89871" containerName="ovn-config" Dec 10 19:19:04 crc kubenswrapper[4828]: E1210 19:19:04.846052 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerName="thanos-sidecar" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.846059 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerName="thanos-sidecar" Dec 10 19:19:04 crc kubenswrapper[4828]: E1210 19:19:04.846091 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerName="prometheus" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.846098 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerName="prometheus" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.846346 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecab69d1-3989-4e69-90f3-b54d4e229283" containerName="mariadb-account-create-update" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.846363 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb2e9f6f-100c-494a-bbca-44c3ab8fc46c" containerName="mariadb-database-create" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.846372 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerName="thanos-sidecar" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.846384 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a97f1c5-2d06-4e5c-8b51-8d0a77a89871" containerName="ovn-config" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.846409 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerName="prometheus" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.846419 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d31859-79eb-4314-9644-455cc33ddacd" containerName="swift-ring-rebalance" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.846435 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" containerName="config-reloader" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.856116 4828 scope.go:117] "RemoveContainer" containerID="1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.857064 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.857414 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.862682 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.862841 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.863015 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.863270 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.863415 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.864626 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-jbqfm" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.869771 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.893576 4828 scope.go:117] "RemoveContainer" containerID="20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4" Dec 10 19:19:04 crc kubenswrapper[4828]: E1210 19:19:04.894354 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4\": container with ID starting with 20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4 not found: ID does not exist" containerID="20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.894396 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4"} err="failed to get container status \"20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4\": rpc error: code = NotFound desc = could not find container \"20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4\": container with ID starting with 20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4 not found: ID does not exist" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.894423 4828 scope.go:117] "RemoveContainer" containerID="b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f" Dec 10 19:19:04 crc kubenswrapper[4828]: E1210 19:19:04.894886 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f\": container with ID starting with b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f not found: ID does not exist" containerID="b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.894940 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f"} err="failed to get container status \"b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f\": rpc error: code = NotFound desc = could not find container \"b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f\": container with ID starting with b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f not found: ID does not exist" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.894974 4828 scope.go:117] "RemoveContainer" containerID="93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b" Dec 10 19:19:04 crc kubenswrapper[4828]: E1210 19:19:04.895306 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b\": container with ID starting with 93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b not found: ID does not exist" containerID="93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.895333 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b"} err="failed to get container status \"93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b\": rpc error: code = NotFound desc = could not find container \"93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b\": container with ID starting with 93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b not found: ID does not exist" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.895348 4828 scope.go:117] "RemoveContainer" containerID="1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3" Dec 10 19:19:04 crc kubenswrapper[4828]: E1210 19:19:04.895638 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3\": container with ID starting with 1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3 not found: ID does not exist" containerID="1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.895677 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3"} err="failed to get container status \"1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3\": rpc error: code = NotFound desc = could not find container \"1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3\": container with ID starting with 1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3 not found: ID does not exist" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.895706 4828 scope.go:117] "RemoveContainer" containerID="20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.896217 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4"} err="failed to get container status \"20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4\": rpc error: code = NotFound desc = could not find container \"20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4\": container with ID starting with 20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4 not found: ID does not exist" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.896295 4828 scope.go:117] "RemoveContainer" containerID="b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.896707 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f"} err="failed to get container status \"b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f\": rpc error: code = NotFound desc = could not find container \"b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f\": container with ID starting with b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f not found: ID does not exist" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.896726 4828 scope.go:117] "RemoveContainer" containerID="93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.899264 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b"} err="failed to get container status \"93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b\": rpc error: code = NotFound desc = could not find container \"93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b\": container with ID starting with 93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b not found: ID does not exist" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.899290 4828 scope.go:117] "RemoveContainer" containerID="1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.899770 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3"} err="failed to get container status \"1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3\": rpc error: code = NotFound desc = could not find container \"1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3\": container with ID starting with 1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3 not found: ID does not exist" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.899789 4828 scope.go:117] "RemoveContainer" containerID="20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.900221 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4"} err="failed to get container status \"20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4\": rpc error: code = NotFound desc = could not find container \"20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4\": container with ID starting with 20a149c918671bf62e94e606d6406cfe8ee43ebd769f2ab48c2ea01443f963e4 not found: ID does not exist" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.900239 4828 scope.go:117] "RemoveContainer" containerID="b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.900573 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f"} err="failed to get container status \"b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f\": rpc error: code = NotFound desc = could not find container \"b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f\": container with ID starting with b1d1bae2b00a61f22344dbf0c0ada3e6dc667080c6d2372e4a3130b9843a235f not found: ID does not exist" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.900595 4828 scope.go:117] "RemoveContainer" containerID="93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.900923 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b"} err="failed to get container status \"93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b\": rpc error: code = NotFound desc = could not find container \"93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b\": container with ID starting with 93338a2faa337e4bd7375e92fe20f49635bab6e0996cad63daf41437c11d0f6b not found: ID does not exist" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.900946 4828 scope.go:117] "RemoveContainer" containerID="1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3" Dec 10 19:19:04 crc kubenswrapper[4828]: I1210 19:19:04.901338 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3"} err="failed to get container status \"1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3\": rpc error: code = NotFound desc = could not find container \"1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3\": container with ID starting with 1266b6018c1e79eb03e64902776a4af04d0d90de6d54f82030df93954a6d62c3 not found: ID does not exist" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.005610 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.005676 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.005726 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.005747 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhh58\" (UniqueName: \"kubernetes.io/projected/5d6e68b4-1320-4111-a47d-51efcfab16b3-kube-api-access-nhh58\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.005834 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.005860 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-config\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.005956 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.006106 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.006156 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5d6e68b4-1320-4111-a47d-51efcfab16b3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.006219 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5d6e68b4-1320-4111-a47d-51efcfab16b3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.006251 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5d6e68b4-1320-4111-a47d-51efcfab16b3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.108984 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.109050 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-config\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.109083 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.109160 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.109194 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5d6e68b4-1320-4111-a47d-51efcfab16b3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.109244 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5d6e68b4-1320-4111-a47d-51efcfab16b3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.109268 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5d6e68b4-1320-4111-a47d-51efcfab16b3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.109338 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.109374 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.109431 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.109454 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhh58\" (UniqueName: \"kubernetes.io/projected/5d6e68b4-1320-4111-a47d-51efcfab16b3-kube-api-access-nhh58\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.109906 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.111522 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5d6e68b4-1320-4111-a47d-51efcfab16b3-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.113314 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5d6e68b4-1320-4111-a47d-51efcfab16b3-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.114164 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.114239 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.114574 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.115890 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-config\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.117020 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5d6e68b4-1320-4111-a47d-51efcfab16b3-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.118630 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.119896 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d6e68b4-1320-4111-a47d-51efcfab16b3-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.128946 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhh58\" (UniqueName: \"kubernetes.io/projected/5d6e68b4-1320-4111-a47d-51efcfab16b3-kube-api-access-nhh58\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.157064 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"5d6e68b4-1320-4111-a47d-51efcfab16b3\") " pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.179844 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.714341 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 10 19:19:05 crc kubenswrapper[4828]: W1210 19:19:05.724863 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d6e68b4_1320_4111_a47d_51efcfab16b3.slice/crio-0d36f18276b98da8c5a813ee3d0cad7051424bc54b32adc648dcc4f03a5e263b WatchSource:0}: Error finding container 0d36f18276b98da8c5a813ee3d0cad7051424bc54b32adc648dcc4f03a5e263b: Status 404 returned error can't find the container with id 0d36f18276b98da8c5a813ee3d0cad7051424bc54b32adc648dcc4f03a5e263b Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.776381 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5d6e68b4-1320-4111-a47d-51efcfab16b3","Type":"ContainerStarted","Data":"0d36f18276b98da8c5a813ee3d0cad7051424bc54b32adc648dcc4f03a5e263b"} Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.801401 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a97f1c5-2d06-4e5c-8b51-8d0a77a89871" path="/var/lib/kubelet/pods/1a97f1c5-2d06-4e5c-8b51-8d0a77a89871/volumes" Dec 10 19:19:05 crc kubenswrapper[4828]: I1210 19:19:05.802458 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0" path="/var/lib/kubelet/pods/975dc70d-7e7e-42a6-b5c2-15d2b52bd3d0/volumes" Dec 10 19:19:08 crc kubenswrapper[4828]: I1210 19:19:08.821646 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5d6e68b4-1320-4111-a47d-51efcfab16b3","Type":"ContainerStarted","Data":"ea63c03d4d676bd49e2f0d37b65c739b3492ab810c2744941474a99f66f554d9"} Dec 10 19:19:08 crc kubenswrapper[4828]: I1210 19:19:08.929969 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Dec 10 19:19:08 crc kubenswrapper[4828]: I1210 19:19:08.931618 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 10 19:19:08 crc kubenswrapper[4828]: I1210 19:19:08.933872 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Dec 10 19:19:08 crc kubenswrapper[4828]: I1210 19:19:08.939613 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 10 19:19:09 crc kubenswrapper[4828]: I1210 19:19:09.090222 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fd3d999-4b0d-422c-a7ae-01065d577c38-config-data\") pod \"mysqld-exporter-0\" (UID: \"9fd3d999-4b0d-422c-a7ae-01065d577c38\") " pod="openstack/mysqld-exporter-0" Dec 10 19:19:09 crc kubenswrapper[4828]: I1210 19:19:09.090322 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd3d999-4b0d-422c-a7ae-01065d577c38-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"9fd3d999-4b0d-422c-a7ae-01065d577c38\") " pod="openstack/mysqld-exporter-0" Dec 10 19:19:09 crc kubenswrapper[4828]: I1210 19:19:09.090387 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfw5c\" (UniqueName: \"kubernetes.io/projected/9fd3d999-4b0d-422c-a7ae-01065d577c38-kube-api-access-rfw5c\") pod \"mysqld-exporter-0\" (UID: \"9fd3d999-4b0d-422c-a7ae-01065d577c38\") " pod="openstack/mysqld-exporter-0" Dec 10 19:19:09 crc kubenswrapper[4828]: I1210 19:19:09.194237 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd3d999-4b0d-422c-a7ae-01065d577c38-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"9fd3d999-4b0d-422c-a7ae-01065d577c38\") " pod="openstack/mysqld-exporter-0" Dec 10 19:19:09 crc kubenswrapper[4828]: I1210 19:19:09.194356 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfw5c\" (UniqueName: \"kubernetes.io/projected/9fd3d999-4b0d-422c-a7ae-01065d577c38-kube-api-access-rfw5c\") pod \"mysqld-exporter-0\" (UID: \"9fd3d999-4b0d-422c-a7ae-01065d577c38\") " pod="openstack/mysqld-exporter-0" Dec 10 19:19:09 crc kubenswrapper[4828]: I1210 19:19:09.194618 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fd3d999-4b0d-422c-a7ae-01065d577c38-config-data\") pod \"mysqld-exporter-0\" (UID: \"9fd3d999-4b0d-422c-a7ae-01065d577c38\") " pod="openstack/mysqld-exporter-0" Dec 10 19:19:09 crc kubenswrapper[4828]: I1210 19:19:09.201706 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fd3d999-4b0d-422c-a7ae-01065d577c38-config-data\") pod \"mysqld-exporter-0\" (UID: \"9fd3d999-4b0d-422c-a7ae-01065d577c38\") " pod="openstack/mysqld-exporter-0" Dec 10 19:19:09 crc kubenswrapper[4828]: I1210 19:19:09.203487 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd3d999-4b0d-422c-a7ae-01065d577c38-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"9fd3d999-4b0d-422c-a7ae-01065d577c38\") " pod="openstack/mysqld-exporter-0" Dec 10 19:19:09 crc kubenswrapper[4828]: I1210 19:19:09.212359 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfw5c\" (UniqueName: \"kubernetes.io/projected/9fd3d999-4b0d-422c-a7ae-01065d577c38-kube-api-access-rfw5c\") pod \"mysqld-exporter-0\" (UID: \"9fd3d999-4b0d-422c-a7ae-01065d577c38\") " pod="openstack/mysqld-exporter-0" Dec 10 19:19:09 crc kubenswrapper[4828]: I1210 19:19:09.256754 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 10 19:19:11 crc kubenswrapper[4828]: I1210 19:19:11.443345 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:19:11 crc kubenswrapper[4828]: I1210 19:19:11.450449 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e29bd910-2077-4ebe-8028-e810061fb704-etc-swift\") pod \"swift-storage-0\" (UID: \"e29bd910-2077-4ebe-8028-e810061fb704\") " pod="openstack/swift-storage-0" Dec 10 19:19:11 crc kubenswrapper[4828]: I1210 19:19:11.707192 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 10 19:19:12 crc kubenswrapper[4828]: I1210 19:19:12.569009 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:19:12 crc kubenswrapper[4828]: I1210 19:19:12.657725 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.256585 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-bbtgm"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.258413 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-bbtgm" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.267641 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-bbtgm"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.399695 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mcvs\" (UniqueName: \"kubernetes.io/projected/a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774-kube-api-access-4mcvs\") pod \"heat-db-create-bbtgm\" (UID: \"a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774\") " pod="openstack/heat-db-create-bbtgm" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.400178 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774-operator-scripts\") pod \"heat-db-create-bbtgm\" (UID: \"a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774\") " pod="openstack/heat-db-create-bbtgm" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.456965 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-bmrgx"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.458618 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bmrgx" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.467466 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-1a9d-account-create-update-ql2pg"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.468885 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-1a9d-account-create-update-ql2pg" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.472639 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.501458 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mcvs\" (UniqueName: \"kubernetes.io/projected/a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774-kube-api-access-4mcvs\") pod \"heat-db-create-bbtgm\" (UID: \"a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774\") " pod="openstack/heat-db-create-bbtgm" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.501605 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/489d1351-3bec-4909-a5dd-7f5e14f280d6-operator-scripts\") pod \"cinder-db-create-bmrgx\" (UID: \"489d1351-3bec-4909-a5dd-7f5e14f280d6\") " pod="openstack/cinder-db-create-bmrgx" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.501638 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774-operator-scripts\") pod \"heat-db-create-bbtgm\" (UID: \"a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774\") " pod="openstack/heat-db-create-bbtgm" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.501683 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8tks\" (UniqueName: \"kubernetes.io/projected/489d1351-3bec-4909-a5dd-7f5e14f280d6-kube-api-access-j8tks\") pod \"cinder-db-create-bmrgx\" (UID: \"489d1351-3bec-4909-a5dd-7f5e14f280d6\") " pod="openstack/cinder-db-create-bmrgx" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.502534 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774-operator-scripts\") pod \"heat-db-create-bbtgm\" (UID: \"a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774\") " pod="openstack/heat-db-create-bbtgm" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.502595 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-bmrgx"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.513930 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-1a9d-account-create-update-ql2pg"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.608032 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/489d1351-3bec-4909-a5dd-7f5e14f280d6-operator-scripts\") pod \"cinder-db-create-bmrgx\" (UID: \"489d1351-3bec-4909-a5dd-7f5e14f280d6\") " pod="openstack/cinder-db-create-bmrgx" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.608392 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8tks\" (UniqueName: \"kubernetes.io/projected/489d1351-3bec-4909-a5dd-7f5e14f280d6-kube-api-access-j8tks\") pod \"cinder-db-create-bmrgx\" (UID: \"489d1351-3bec-4909-a5dd-7f5e14f280d6\") " pod="openstack/cinder-db-create-bmrgx" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.609575 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/489d1351-3bec-4909-a5dd-7f5e14f280d6-operator-scripts\") pod \"cinder-db-create-bmrgx\" (UID: \"489d1351-3bec-4909-a5dd-7f5e14f280d6\") " pod="openstack/cinder-db-create-bmrgx" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.611661 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mcvs\" (UniqueName: \"kubernetes.io/projected/a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774-kube-api-access-4mcvs\") pod \"heat-db-create-bbtgm\" (UID: \"a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774\") " pod="openstack/heat-db-create-bbtgm" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.635089 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-b2rhf"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.636478 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-b2rhf" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.640172 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.640552 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.640909 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.643750 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-b2rhf"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.645683 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-c922f" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.648948 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8tks\" (UniqueName: \"kubernetes.io/projected/489d1351-3bec-4909-a5dd-7f5e14f280d6-kube-api-access-j8tks\") pod \"cinder-db-create-bmrgx\" (UID: \"489d1351-3bec-4909-a5dd-7f5e14f280d6\") " pod="openstack/cinder-db-create-bmrgx" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.673038 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-4btg4"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.675244 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4btg4" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.692625 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4btg4"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.710633 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfzbk\" (UniqueName: \"kubernetes.io/projected/37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3-kube-api-access-qfzbk\") pod \"heat-1a9d-account-create-update-ql2pg\" (UID: \"37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3\") " pod="openstack/heat-1a9d-account-create-update-ql2pg" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.710868 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3-operator-scripts\") pod \"heat-1a9d-account-create-update-ql2pg\" (UID: \"37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3\") " pod="openstack/heat-1a9d-account-create-update-ql2pg" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.715559 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-d565-account-create-update-vwzdt"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.717406 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d565-account-create-update-vwzdt" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.722230 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.748696 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d565-account-create-update-vwzdt"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.774012 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-zkd8f"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.776934 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zkd8f" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.784421 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-zkd8f"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.797132 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bmrgx" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.813300 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611d78ee-8203-4a4f-bb01-eedd92ab42fd-combined-ca-bundle\") pod \"keystone-db-sync-b2rhf\" (UID: \"611d78ee-8203-4a4f-bb01-eedd92ab42fd\") " pod="openstack/keystone-db-sync-b2rhf" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.813536 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3-operator-scripts\") pod \"heat-1a9d-account-create-update-ql2pg\" (UID: \"37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3\") " pod="openstack/heat-1a9d-account-create-update-ql2pg" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.813609 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca06dcbb-f357-454e-afd5-7bd5b42fb6ed-operator-scripts\") pod \"barbican-db-create-4btg4\" (UID: \"ca06dcbb-f357-454e-afd5-7bd5b42fb6ed\") " pod="openstack/barbican-db-create-4btg4" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.813678 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrwqf\" (UniqueName: \"kubernetes.io/projected/ca06dcbb-f357-454e-afd5-7bd5b42fb6ed-kube-api-access-zrwqf\") pod \"barbican-db-create-4btg4\" (UID: \"ca06dcbb-f357-454e-afd5-7bd5b42fb6ed\") " pod="openstack/barbican-db-create-4btg4" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.813720 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/611d78ee-8203-4a4f-bb01-eedd92ab42fd-config-data\") pod \"keystone-db-sync-b2rhf\" (UID: \"611d78ee-8203-4a4f-bb01-eedd92ab42fd\") " pod="openstack/keystone-db-sync-b2rhf" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.813743 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv9q7\" (UniqueName: \"kubernetes.io/projected/611d78ee-8203-4a4f-bb01-eedd92ab42fd-kube-api-access-nv9q7\") pod \"keystone-db-sync-b2rhf\" (UID: \"611d78ee-8203-4a4f-bb01-eedd92ab42fd\") " pod="openstack/keystone-db-sync-b2rhf" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.813905 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfzbk\" (UniqueName: \"kubernetes.io/projected/37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3-kube-api-access-qfzbk\") pod \"heat-1a9d-account-create-update-ql2pg\" (UID: \"37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3\") " pod="openstack/heat-1a9d-account-create-update-ql2pg" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.816083 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3-operator-scripts\") pod \"heat-1a9d-account-create-update-ql2pg\" (UID: \"37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3\") " pod="openstack/heat-1a9d-account-create-update-ql2pg" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.816716 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-30e8-account-create-update-5zwls"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.818482 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-30e8-account-create-update-5zwls" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.820939 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.828503 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-30e8-account-create-update-5zwls"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.838825 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfzbk\" (UniqueName: \"kubernetes.io/projected/37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3-kube-api-access-qfzbk\") pod \"heat-1a9d-account-create-update-ql2pg\" (UID: \"37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3\") " pod="openstack/heat-1a9d-account-create-update-ql2pg" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.896847 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-bbtgm" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.913445 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-ae3c-account-create-update-v6g92"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.914995 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ae3c-account-create-update-v6g92" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.916174 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/515a9b7f-6551-4a40-bd25-6c851129a6d0-operator-scripts\") pod \"neutron-db-create-zkd8f\" (UID: \"515a9b7f-6551-4a40-bd25-6c851129a6d0\") " pod="openstack/neutron-db-create-zkd8f" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.916228 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cea49d65-9ca3-41a3-b970-25b0ccaac138-operator-scripts\") pod \"barbican-d565-account-create-update-vwzdt\" (UID: \"cea49d65-9ca3-41a3-b970-25b0ccaac138\") " pod="openstack/barbican-d565-account-create-update-vwzdt" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.916273 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611d78ee-8203-4a4f-bb01-eedd92ab42fd-combined-ca-bundle\") pod \"keystone-db-sync-b2rhf\" (UID: \"611d78ee-8203-4a4f-bb01-eedd92ab42fd\") " pod="openstack/keystone-db-sync-b2rhf" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.916343 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlhzg\" (UniqueName: \"kubernetes.io/projected/515a9b7f-6551-4a40-bd25-6c851129a6d0-kube-api-access-rlhzg\") pod \"neutron-db-create-zkd8f\" (UID: \"515a9b7f-6551-4a40-bd25-6c851129a6d0\") " pod="openstack/neutron-db-create-zkd8f" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.916378 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca06dcbb-f357-454e-afd5-7bd5b42fb6ed-operator-scripts\") pod \"barbican-db-create-4btg4\" (UID: \"ca06dcbb-f357-454e-afd5-7bd5b42fb6ed\") " pod="openstack/barbican-db-create-4btg4" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.916408 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrwqf\" (UniqueName: \"kubernetes.io/projected/ca06dcbb-f357-454e-afd5-7bd5b42fb6ed-kube-api-access-zrwqf\") pod \"barbican-db-create-4btg4\" (UID: \"ca06dcbb-f357-454e-afd5-7bd5b42fb6ed\") " pod="openstack/barbican-db-create-4btg4" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.916430 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/611d78ee-8203-4a4f-bb01-eedd92ab42fd-config-data\") pod \"keystone-db-sync-b2rhf\" (UID: \"611d78ee-8203-4a4f-bb01-eedd92ab42fd\") " pod="openstack/keystone-db-sync-b2rhf" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.916450 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv9q7\" (UniqueName: \"kubernetes.io/projected/611d78ee-8203-4a4f-bb01-eedd92ab42fd-kube-api-access-nv9q7\") pod \"keystone-db-sync-b2rhf\" (UID: \"611d78ee-8203-4a4f-bb01-eedd92ab42fd\") " pod="openstack/keystone-db-sync-b2rhf" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.916499 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4krs\" (UniqueName: \"kubernetes.io/projected/cea49d65-9ca3-41a3-b970-25b0ccaac138-kube-api-access-m4krs\") pod \"barbican-d565-account-create-update-vwzdt\" (UID: \"cea49d65-9ca3-41a3-b970-25b0ccaac138\") " pod="openstack/barbican-d565-account-create-update-vwzdt" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.918310 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca06dcbb-f357-454e-afd5-7bd5b42fb6ed-operator-scripts\") pod \"barbican-db-create-4btg4\" (UID: \"ca06dcbb-f357-454e-afd5-7bd5b42fb6ed\") " pod="openstack/barbican-db-create-4btg4" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.918667 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.921191 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/611d78ee-8203-4a4f-bb01-eedd92ab42fd-config-data\") pod \"keystone-db-sync-b2rhf\" (UID: \"611d78ee-8203-4a4f-bb01-eedd92ab42fd\") " pod="openstack/keystone-db-sync-b2rhf" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.923685 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611d78ee-8203-4a4f-bb01-eedd92ab42fd-combined-ca-bundle\") pod \"keystone-db-sync-b2rhf\" (UID: \"611d78ee-8203-4a4f-bb01-eedd92ab42fd\") " pod="openstack/keystone-db-sync-b2rhf" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.924483 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ae3c-account-create-update-v6g92"] Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.943307 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv9q7\" (UniqueName: \"kubernetes.io/projected/611d78ee-8203-4a4f-bb01-eedd92ab42fd-kube-api-access-nv9q7\") pod \"keystone-db-sync-b2rhf\" (UID: \"611d78ee-8203-4a4f-bb01-eedd92ab42fd\") " pod="openstack/keystone-db-sync-b2rhf" Dec 10 19:19:14 crc kubenswrapper[4828]: I1210 19:19:14.946462 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrwqf\" (UniqueName: \"kubernetes.io/projected/ca06dcbb-f357-454e-afd5-7bd5b42fb6ed-kube-api-access-zrwqf\") pod \"barbican-db-create-4btg4\" (UID: \"ca06dcbb-f357-454e-afd5-7bd5b42fb6ed\") " pod="openstack/barbican-db-create-4btg4" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.009370 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-b2rhf" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.021912 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20f5475c-6d07-4be3-96f5-fed87f4ad62a-operator-scripts\") pod \"cinder-30e8-account-create-update-5zwls\" (UID: \"20f5475c-6d07-4be3-96f5-fed87f4ad62a\") " pod="openstack/cinder-30e8-account-create-update-5zwls" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.021965 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdfzk\" (UniqueName: \"kubernetes.io/projected/43f54580-b31f-45db-b046-3fb45235c384-kube-api-access-wdfzk\") pod \"neutron-ae3c-account-create-update-v6g92\" (UID: \"43f54580-b31f-45db-b046-3fb45235c384\") " pod="openstack/neutron-ae3c-account-create-update-v6g92" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.022019 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4krs\" (UniqueName: \"kubernetes.io/projected/cea49d65-9ca3-41a3-b970-25b0ccaac138-kube-api-access-m4krs\") pod \"barbican-d565-account-create-update-vwzdt\" (UID: \"cea49d65-9ca3-41a3-b970-25b0ccaac138\") " pod="openstack/barbican-d565-account-create-update-vwzdt" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.022106 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/515a9b7f-6551-4a40-bd25-6c851129a6d0-operator-scripts\") pod \"neutron-db-create-zkd8f\" (UID: \"515a9b7f-6551-4a40-bd25-6c851129a6d0\") " pod="openstack/neutron-db-create-zkd8f" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.022140 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43f54580-b31f-45db-b046-3fb45235c384-operator-scripts\") pod \"neutron-ae3c-account-create-update-v6g92\" (UID: \"43f54580-b31f-45db-b046-3fb45235c384\") " pod="openstack/neutron-ae3c-account-create-update-v6g92" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.022198 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cea49d65-9ca3-41a3-b970-25b0ccaac138-operator-scripts\") pod \"barbican-d565-account-create-update-vwzdt\" (UID: \"cea49d65-9ca3-41a3-b970-25b0ccaac138\") " pod="openstack/barbican-d565-account-create-update-vwzdt" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.022267 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlhzg\" (UniqueName: \"kubernetes.io/projected/515a9b7f-6551-4a40-bd25-6c851129a6d0-kube-api-access-rlhzg\") pod \"neutron-db-create-zkd8f\" (UID: \"515a9b7f-6551-4a40-bd25-6c851129a6d0\") " pod="openstack/neutron-db-create-zkd8f" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.022323 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmsld\" (UniqueName: \"kubernetes.io/projected/20f5475c-6d07-4be3-96f5-fed87f4ad62a-kube-api-access-dmsld\") pod \"cinder-30e8-account-create-update-5zwls\" (UID: \"20f5475c-6d07-4be3-96f5-fed87f4ad62a\") " pod="openstack/cinder-30e8-account-create-update-5zwls" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.023413 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/515a9b7f-6551-4a40-bd25-6c851129a6d0-operator-scripts\") pod \"neutron-db-create-zkd8f\" (UID: \"515a9b7f-6551-4a40-bd25-6c851129a6d0\") " pod="openstack/neutron-db-create-zkd8f" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.023940 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cea49d65-9ca3-41a3-b970-25b0ccaac138-operator-scripts\") pod \"barbican-d565-account-create-update-vwzdt\" (UID: \"cea49d65-9ca3-41a3-b970-25b0ccaac138\") " pod="openstack/barbican-d565-account-create-update-vwzdt" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.031533 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4btg4" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.047729 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlhzg\" (UniqueName: \"kubernetes.io/projected/515a9b7f-6551-4a40-bd25-6c851129a6d0-kube-api-access-rlhzg\") pod \"neutron-db-create-zkd8f\" (UID: \"515a9b7f-6551-4a40-bd25-6c851129a6d0\") " pod="openstack/neutron-db-create-zkd8f" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.059071 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4krs\" (UniqueName: \"kubernetes.io/projected/cea49d65-9ca3-41a3-b970-25b0ccaac138-kube-api-access-m4krs\") pod \"barbican-d565-account-create-update-vwzdt\" (UID: \"cea49d65-9ca3-41a3-b970-25b0ccaac138\") " pod="openstack/barbican-d565-account-create-update-vwzdt" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.104848 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zkd8f" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.112741 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-1a9d-account-create-update-ql2pg" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.124478 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmsld\" (UniqueName: \"kubernetes.io/projected/20f5475c-6d07-4be3-96f5-fed87f4ad62a-kube-api-access-dmsld\") pod \"cinder-30e8-account-create-update-5zwls\" (UID: \"20f5475c-6d07-4be3-96f5-fed87f4ad62a\") " pod="openstack/cinder-30e8-account-create-update-5zwls" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.124605 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20f5475c-6d07-4be3-96f5-fed87f4ad62a-operator-scripts\") pod \"cinder-30e8-account-create-update-5zwls\" (UID: \"20f5475c-6d07-4be3-96f5-fed87f4ad62a\") " pod="openstack/cinder-30e8-account-create-update-5zwls" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.125341 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20f5475c-6d07-4be3-96f5-fed87f4ad62a-operator-scripts\") pod \"cinder-30e8-account-create-update-5zwls\" (UID: \"20f5475c-6d07-4be3-96f5-fed87f4ad62a\") " pod="openstack/cinder-30e8-account-create-update-5zwls" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.124634 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdfzk\" (UniqueName: \"kubernetes.io/projected/43f54580-b31f-45db-b046-3fb45235c384-kube-api-access-wdfzk\") pod \"neutron-ae3c-account-create-update-v6g92\" (UID: \"43f54580-b31f-45db-b046-3fb45235c384\") " pod="openstack/neutron-ae3c-account-create-update-v6g92" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.125529 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43f54580-b31f-45db-b046-3fb45235c384-operator-scripts\") pod \"neutron-ae3c-account-create-update-v6g92\" (UID: \"43f54580-b31f-45db-b046-3fb45235c384\") " pod="openstack/neutron-ae3c-account-create-update-v6g92" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.126104 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43f54580-b31f-45db-b046-3fb45235c384-operator-scripts\") pod \"neutron-ae3c-account-create-update-v6g92\" (UID: \"43f54580-b31f-45db-b046-3fb45235c384\") " pod="openstack/neutron-ae3c-account-create-update-v6g92" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.142849 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmsld\" (UniqueName: \"kubernetes.io/projected/20f5475c-6d07-4be3-96f5-fed87f4ad62a-kube-api-access-dmsld\") pod \"cinder-30e8-account-create-update-5zwls\" (UID: \"20f5475c-6d07-4be3-96f5-fed87f4ad62a\") " pod="openstack/cinder-30e8-account-create-update-5zwls" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.148342 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdfzk\" (UniqueName: \"kubernetes.io/projected/43f54580-b31f-45db-b046-3fb45235c384-kube-api-access-wdfzk\") pod \"neutron-ae3c-account-create-update-v6g92\" (UID: \"43f54580-b31f-45db-b046-3fb45235c384\") " pod="openstack/neutron-ae3c-account-create-update-v6g92" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.179080 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-30e8-account-create-update-5zwls" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.339838 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ae3c-account-create-update-v6g92" Dec 10 19:19:15 crc kubenswrapper[4828]: I1210 19:19:15.339844 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d565-account-create-update-vwzdt" Dec 10 19:19:16 crc kubenswrapper[4828]: I1210 19:19:16.920885 4828 generic.go:334] "Generic (PLEG): container finished" podID="5d6e68b4-1320-4111-a47d-51efcfab16b3" containerID="ea63c03d4d676bd49e2f0d37b65c739b3492ab810c2744941474a99f66f554d9" exitCode=0 Dec 10 19:19:16 crc kubenswrapper[4828]: I1210 19:19:16.920955 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5d6e68b4-1320-4111-a47d-51efcfab16b3","Type":"ContainerDied","Data":"ea63c03d4d676bd49e2f0d37b65c739b3492ab810c2744941474a99f66f554d9"} Dec 10 19:19:17 crc kubenswrapper[4828]: I1210 19:19:17.649516 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-bmrgx"] Dec 10 19:19:17 crc kubenswrapper[4828]: W1210 19:19:17.740266 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod489d1351_3bec_4909_a5dd_7f5e14f280d6.slice/crio-a4363c2ccdd0fefc7a6936e9802ce2abdb76729d00b65e599c651fbe3a830fe4 WatchSource:0}: Error finding container a4363c2ccdd0fefc7a6936e9802ce2abdb76729d00b65e599c651fbe3a830fe4: Status 404 returned error can't find the container with id a4363c2ccdd0fefc7a6936e9802ce2abdb76729d00b65e599c651fbe3a830fe4 Dec 10 19:19:17 crc kubenswrapper[4828]: I1210 19:19:17.940027 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5d6e68b4-1320-4111-a47d-51efcfab16b3","Type":"ContainerStarted","Data":"05a951086682abd9e1e3e1eade86d5d6440daa7d53f94295b99db7dd04222b53"} Dec 10 19:19:17 crc kubenswrapper[4828]: I1210 19:19:17.943102 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bmrgx" event={"ID":"489d1351-3bec-4909-a5dd-7f5e14f280d6","Type":"ContainerStarted","Data":"a4363c2ccdd0fefc7a6936e9802ce2abdb76729d00b65e599c651fbe3a830fe4"} Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.257336 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 10 19:19:18 crc kubenswrapper[4828]: W1210 19:19:18.260696 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fd3d999_4b0d_422c_a7ae_01065d577c38.slice/crio-546859861b4dd77704052864b69c1da3edffa0082c086146dca7634985726909 WatchSource:0}: Error finding container 546859861b4dd77704052864b69c1da3edffa0082c086146dca7634985726909: Status 404 returned error can't find the container with id 546859861b4dd77704052864b69c1da3edffa0082c086146dca7634985726909 Dec 10 19:19:18 crc kubenswrapper[4828]: W1210 19:19:18.266283 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43f54580_b31f_45db_b046_3fb45235c384.slice/crio-0a735ceed02220bd65b1d7ffded2f84c12ff35a3bca815826bea3acae99f6ab8 WatchSource:0}: Error finding container 0a735ceed02220bd65b1d7ffded2f84c12ff35a3bca815826bea3acae99f6ab8: Status 404 returned error can't find the container with id 0a735ceed02220bd65b1d7ffded2f84c12ff35a3bca815826bea3acae99f6ab8 Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.273205 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ae3c-account-create-update-v6g92"] Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.284101 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-zkd8f"] Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.501088 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-1a9d-account-create-update-ql2pg"] Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.526767 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d565-account-create-update-vwzdt"] Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.546629 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-30e8-account-create-update-5zwls"] Dec 10 19:19:18 crc kubenswrapper[4828]: W1210 19:19:18.556845 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca06dcbb_f357_454e_afd5_7bd5b42fb6ed.slice/crio-0b062ed37673ad0899d9278be3fea7f9e350637a3e3480c5d1672a2b0a3b9a08 WatchSource:0}: Error finding container 0b062ed37673ad0899d9278be3fea7f9e350637a3e3480c5d1672a2b0a3b9a08: Status 404 returned error can't find the container with id 0b062ed37673ad0899d9278be3fea7f9e350637a3e3480c5d1672a2b0a3b9a08 Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.568438 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-b2rhf"] Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.581648 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4btg4"] Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.607020 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.625788 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-bbtgm"] Dec 10 19:19:18 crc kubenswrapper[4828]: W1210 19:19:18.627272 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode29bd910_2077_4ebe_8028_e810061fb704.slice/crio-ea01835def7098adf9d50080e0a8d136b9cb15a13268ff1e7e8826dfa75a6eee WatchSource:0}: Error finding container ea01835def7098adf9d50080e0a8d136b9cb15a13268ff1e7e8826dfa75a6eee: Status 404 returned error can't find the container with id ea01835def7098adf9d50080e0a8d136b9cb15a13268ff1e7e8826dfa75a6eee Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.955689 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-30e8-account-create-update-5zwls" event={"ID":"20f5475c-6d07-4be3-96f5-fed87f4ad62a","Type":"ContainerStarted","Data":"75cbada4d797e4865c6ff26b2f8810a88e582e6f12177fa6789a1d943d960bb8"} Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.958465 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9rgqj" event={"ID":"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2","Type":"ContainerStarted","Data":"5a669fd9dee5beaa44c3ac23322606f4e2ad87a45eca78cc6e7f974d677c7b6d"} Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.961314 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"9fd3d999-4b0d-422c-a7ae-01065d577c38","Type":"ContainerStarted","Data":"546859861b4dd77704052864b69c1da3edffa0082c086146dca7634985726909"} Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.965631 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4btg4" event={"ID":"ca06dcbb-f357-454e-afd5-7bd5b42fb6ed","Type":"ContainerStarted","Data":"0b062ed37673ad0899d9278be3fea7f9e350637a3e3480c5d1672a2b0a3b9a08"} Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.967087 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ae3c-account-create-update-v6g92" event={"ID":"43f54580-b31f-45db-b046-3fb45235c384","Type":"ContainerStarted","Data":"e3ff992a8c33d465274becc00390c4ea69eb05aa6d8dc109e5e8b2cf8bbf36b7"} Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.967132 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ae3c-account-create-update-v6g92" event={"ID":"43f54580-b31f-45db-b046-3fb45235c384","Type":"ContainerStarted","Data":"0a735ceed02220bd65b1d7ffded2f84c12ff35a3bca815826bea3acae99f6ab8"} Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.969143 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-1a9d-account-create-update-ql2pg" event={"ID":"37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3","Type":"ContainerStarted","Data":"89bf1a217038cec0da206b74444329ac2674155ea21be96652d74c3feddb3d09"} Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.969172 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-1a9d-account-create-update-ql2pg" event={"ID":"37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3","Type":"ContainerStarted","Data":"067a00e41305df1c07cc77071effb7fd118e14897663ef4faeb897df9745f958"} Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.970914 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e29bd910-2077-4ebe-8028-e810061fb704","Type":"ContainerStarted","Data":"ea01835def7098adf9d50080e0a8d136b9cb15a13268ff1e7e8826dfa75a6eee"} Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.977373 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-bbtgm" event={"ID":"a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774","Type":"ContainerStarted","Data":"129f3009b77db2ef5f729d0825fd706f68d16e3e9e29d4ae0153599bb09a6065"} Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.989455 4828 generic.go:334] "Generic (PLEG): container finished" podID="489d1351-3bec-4909-a5dd-7f5e14f280d6" containerID="9bd8703eac224ae862977acd7a3bf7bb03820f910fd20187da8b777b36435ddc" exitCode=0 Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.989577 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bmrgx" event={"ID":"489d1351-3bec-4909-a5dd-7f5e14f280d6","Type":"ContainerDied","Data":"9bd8703eac224ae862977acd7a3bf7bb03820f910fd20187da8b777b36435ddc"} Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.992557 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d565-account-create-update-vwzdt" event={"ID":"cea49d65-9ca3-41a3-b970-25b0ccaac138","Type":"ContainerStarted","Data":"37c0fd1c4a255f54a28b8d30226fe28b08bc7d14419a495ff9439f26a9b4b31e"} Dec 10 19:19:18 crc kubenswrapper[4828]: I1210 19:19:18.995138 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-b2rhf" event={"ID":"611d78ee-8203-4a4f-bb01-eedd92ab42fd","Type":"ContainerStarted","Data":"cb2e663ddf153f81089beb1d3e599338f9f050b824d4df01e4756a19cf4344bd"} Dec 10 19:19:19 crc kubenswrapper[4828]: I1210 19:19:18.998994 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zkd8f" event={"ID":"515a9b7f-6551-4a40-bd25-6c851129a6d0","Type":"ContainerStarted","Data":"d869cd6461234fc218a1eaf175e3888c1eee95293ecc651d9ba024736d5d2c2e"} Dec 10 19:19:19 crc kubenswrapper[4828]: I1210 19:19:18.999038 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zkd8f" event={"ID":"515a9b7f-6551-4a40-bd25-6c851129a6d0","Type":"ContainerStarted","Data":"0aef5b1f7bee92f85c25b3b67be94bb1ac4ec9a5047706d4db8e7abed6a9bb6d"} Dec 10 19:19:19 crc kubenswrapper[4828]: I1210 19:19:19.007656 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-ae3c-account-create-update-v6g92" podStartSLOduration=5.007638325 podStartE2EDuration="5.007638325s" podCreationTimestamp="2025-12-10 19:19:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:19.007039799 +0000 UTC m=+1439.517650824" watchObservedRunningTime="2025-12-10 19:19:19.007638325 +0000 UTC m=+1439.518249330" Dec 10 19:19:19 crc kubenswrapper[4828]: I1210 19:19:19.018144 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-9rgqj" podStartSLOduration=3.141531656 podStartE2EDuration="18.018120021s" podCreationTimestamp="2025-12-10 19:19:01 +0000 UTC" firstStartedPulling="2025-12-10 19:19:02.588469026 +0000 UTC m=+1423.099080031" lastFinishedPulling="2025-12-10 19:19:17.465057391 +0000 UTC m=+1437.975668396" observedRunningTime="2025-12-10 19:19:18.985899292 +0000 UTC m=+1439.496510287" watchObservedRunningTime="2025-12-10 19:19:19.018120021 +0000 UTC m=+1439.528731026" Dec 10 19:19:19 crc kubenswrapper[4828]: I1210 19:19:19.046919 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-1a9d-account-create-update-ql2pg" podStartSLOduration=5.046900649 podStartE2EDuration="5.046900649s" podCreationTimestamp="2025-12-10 19:19:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:19.035253892 +0000 UTC m=+1439.545864897" watchObservedRunningTime="2025-12-10 19:19:19.046900649 +0000 UTC m=+1439.557511644" Dec 10 19:19:19 crc kubenswrapper[4828]: I1210 19:19:19.070146 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-zkd8f" podStartSLOduration=5.07012715 podStartE2EDuration="5.07012715s" podCreationTimestamp="2025-12-10 19:19:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:19.063052564 +0000 UTC m=+1439.573663569" watchObservedRunningTime="2025-12-10 19:19:19.07012715 +0000 UTC m=+1439.580738155" Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.009491 4828 generic.go:334] "Generic (PLEG): container finished" podID="a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774" containerID="5697f45ea9e4d482df1cb71f59c4405623f3d5daa5661c7ac75622b2e1e2f4e7" exitCode=0 Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.009522 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-bbtgm" event={"ID":"a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774","Type":"ContainerDied","Data":"5697f45ea9e4d482df1cb71f59c4405623f3d5daa5661c7ac75622b2e1e2f4e7"} Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.012078 4828 generic.go:334] "Generic (PLEG): container finished" podID="cea49d65-9ca3-41a3-b970-25b0ccaac138" containerID="bed6417b33c47d4233837e21e57c382f5464419c78faf437b09a3367f3779dee" exitCode=0 Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.012111 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d565-account-create-update-vwzdt" event={"ID":"cea49d65-9ca3-41a3-b970-25b0ccaac138","Type":"ContainerDied","Data":"bed6417b33c47d4233837e21e57c382f5464419c78faf437b09a3367f3779dee"} Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.013881 4828 generic.go:334] "Generic (PLEG): container finished" podID="515a9b7f-6551-4a40-bd25-6c851129a6d0" containerID="d869cd6461234fc218a1eaf175e3888c1eee95293ecc651d9ba024736d5d2c2e" exitCode=0 Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.014021 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zkd8f" event={"ID":"515a9b7f-6551-4a40-bd25-6c851129a6d0","Type":"ContainerDied","Data":"d869cd6461234fc218a1eaf175e3888c1eee95293ecc651d9ba024736d5d2c2e"} Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.015651 4828 generic.go:334] "Generic (PLEG): container finished" podID="ca06dcbb-f357-454e-afd5-7bd5b42fb6ed" containerID="ec108eeff2ebf2be4a5e63b9f06ab16ac0c30fb65edf59a4dbf7f0bb24d8cf97" exitCode=0 Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.015700 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4btg4" event={"ID":"ca06dcbb-f357-454e-afd5-7bd5b42fb6ed","Type":"ContainerDied","Data":"ec108eeff2ebf2be4a5e63b9f06ab16ac0c30fb65edf59a4dbf7f0bb24d8cf97"} Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.017301 4828 generic.go:334] "Generic (PLEG): container finished" podID="37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3" containerID="89bf1a217038cec0da206b74444329ac2674155ea21be96652d74c3feddb3d09" exitCode=0 Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.017352 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-1a9d-account-create-update-ql2pg" event={"ID":"37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3","Type":"ContainerDied","Data":"89bf1a217038cec0da206b74444329ac2674155ea21be96652d74c3feddb3d09"} Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.019044 4828 generic.go:334] "Generic (PLEG): container finished" podID="43f54580-b31f-45db-b046-3fb45235c384" containerID="e3ff992a8c33d465274becc00390c4ea69eb05aa6d8dc109e5e8b2cf8bbf36b7" exitCode=0 Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.019098 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ae3c-account-create-update-v6g92" event={"ID":"43f54580-b31f-45db-b046-3fb45235c384","Type":"ContainerDied","Data":"e3ff992a8c33d465274becc00390c4ea69eb05aa6d8dc109e5e8b2cf8bbf36b7"} Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.025285 4828 generic.go:334] "Generic (PLEG): container finished" podID="20f5475c-6d07-4be3-96f5-fed87f4ad62a" containerID="48aff4c4f9de4966bd0b4dd353d7c334a79534b9bba9d82347bb061c5ff4b2b1" exitCode=0 Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.025515 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-30e8-account-create-update-5zwls" event={"ID":"20f5475c-6d07-4be3-96f5-fed87f4ad62a","Type":"ContainerDied","Data":"48aff4c4f9de4966bd0b4dd353d7c334a79534b9bba9d82347bb061c5ff4b2b1"} Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.469439 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bmrgx" Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.654989 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8tks\" (UniqueName: \"kubernetes.io/projected/489d1351-3bec-4909-a5dd-7f5e14f280d6-kube-api-access-j8tks\") pod \"489d1351-3bec-4909-a5dd-7f5e14f280d6\" (UID: \"489d1351-3bec-4909-a5dd-7f5e14f280d6\") " Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.655097 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/489d1351-3bec-4909-a5dd-7f5e14f280d6-operator-scripts\") pod \"489d1351-3bec-4909-a5dd-7f5e14f280d6\" (UID: \"489d1351-3bec-4909-a5dd-7f5e14f280d6\") " Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.655627 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/489d1351-3bec-4909-a5dd-7f5e14f280d6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "489d1351-3bec-4909-a5dd-7f5e14f280d6" (UID: "489d1351-3bec-4909-a5dd-7f5e14f280d6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.660660 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/489d1351-3bec-4909-a5dd-7f5e14f280d6-kube-api-access-j8tks" (OuterVolumeSpecName: "kube-api-access-j8tks") pod "489d1351-3bec-4909-a5dd-7f5e14f280d6" (UID: "489d1351-3bec-4909-a5dd-7f5e14f280d6"). InnerVolumeSpecName "kube-api-access-j8tks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.757444 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8tks\" (UniqueName: \"kubernetes.io/projected/489d1351-3bec-4909-a5dd-7f5e14f280d6-kube-api-access-j8tks\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:20 crc kubenswrapper[4828]: I1210 19:19:20.757749 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/489d1351-3bec-4909-a5dd-7f5e14f280d6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:21 crc kubenswrapper[4828]: I1210 19:19:21.037097 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bmrgx" event={"ID":"489d1351-3bec-4909-a5dd-7f5e14f280d6","Type":"ContainerDied","Data":"a4363c2ccdd0fefc7a6936e9802ce2abdb76729d00b65e599c651fbe3a830fe4"} Dec 10 19:19:21 crc kubenswrapper[4828]: I1210 19:19:21.037347 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4363c2ccdd0fefc7a6936e9802ce2abdb76729d00b65e599c651fbe3a830fe4" Dec 10 19:19:21 crc kubenswrapper[4828]: I1210 19:19:21.037125 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bmrgx" Dec 10 19:19:21 crc kubenswrapper[4828]: I1210 19:19:21.039297 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5d6e68b4-1320-4111-a47d-51efcfab16b3","Type":"ContainerStarted","Data":"2ffc8bacfd81e30b589d0c0dd872e916d98ce4d7167c192670d0957e96fd31b3"} Dec 10 19:19:21 crc kubenswrapper[4828]: I1210 19:19:21.945770 4828 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podd2630e12-86b3-45d3-84c0-dd50f5cf8c15"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podd2630e12-86b3-45d3-84c0-dd50f5cf8c15] : Timed out while waiting for systemd to remove kubepods-besteffort-podd2630e12_86b3_45d3_84c0_dd50f5cf8c15.slice" Dec 10 19:19:21 crc kubenswrapper[4828]: E1210 19:19:21.946119 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort podd2630e12-86b3-45d3-84c0-dd50f5cf8c15] : unable to destroy cgroup paths for cgroup [kubepods besteffort podd2630e12-86b3-45d3-84c0-dd50f5cf8c15] : Timed out while waiting for systemd to remove kubepods-besteffort-podd2630e12_86b3_45d3_84c0_dd50f5cf8c15.slice" pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" podUID="d2630e12-86b3-45d3-84c0-dd50f5cf8c15" Dec 10 19:19:22 crc kubenswrapper[4828]: I1210 19:19:22.049210 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-4cpx8" Dec 10 19:19:22 crc kubenswrapper[4828]: I1210 19:19:22.079842 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4cpx8"] Dec 10 19:19:22 crc kubenswrapper[4828]: I1210 19:19:22.091507 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-4cpx8"] Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.400251 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zkd8f" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.434400 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-1a9d-account-create-update-ql2pg" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.443183 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-bbtgm" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.459403 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d565-account-create-update-vwzdt" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.484912 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4btg4" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.494069 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ae3c-account-create-update-v6g92" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.503844 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-30e8-account-create-update-5zwls" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.514521 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlhzg\" (UniqueName: \"kubernetes.io/projected/515a9b7f-6551-4a40-bd25-6c851129a6d0-kube-api-access-rlhzg\") pod \"515a9b7f-6551-4a40-bd25-6c851129a6d0\" (UID: \"515a9b7f-6551-4a40-bd25-6c851129a6d0\") " Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.514698 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/515a9b7f-6551-4a40-bd25-6c851129a6d0-operator-scripts\") pod \"515a9b7f-6551-4a40-bd25-6c851129a6d0\" (UID: \"515a9b7f-6551-4a40-bd25-6c851129a6d0\") " Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.516497 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/515a9b7f-6551-4a40-bd25-6c851129a6d0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "515a9b7f-6551-4a40-bd25-6c851129a6d0" (UID: "515a9b7f-6551-4a40-bd25-6c851129a6d0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.525487 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/515a9b7f-6551-4a40-bd25-6c851129a6d0-kube-api-access-rlhzg" (OuterVolumeSpecName: "kube-api-access-rlhzg") pod "515a9b7f-6551-4a40-bd25-6c851129a6d0" (UID: "515a9b7f-6551-4a40-bd25-6c851129a6d0"). InnerVolumeSpecName "kube-api-access-rlhzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.616761 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20f5475c-6d07-4be3-96f5-fed87f4ad62a-operator-scripts\") pod \"20f5475c-6d07-4be3-96f5-fed87f4ad62a\" (UID: \"20f5475c-6d07-4be3-96f5-fed87f4ad62a\") " Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.617062 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4krs\" (UniqueName: \"kubernetes.io/projected/cea49d65-9ca3-41a3-b970-25b0ccaac138-kube-api-access-m4krs\") pod \"cea49d65-9ca3-41a3-b970-25b0ccaac138\" (UID: \"cea49d65-9ca3-41a3-b970-25b0ccaac138\") " Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.617213 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca06dcbb-f357-454e-afd5-7bd5b42fb6ed-operator-scripts\") pod \"ca06dcbb-f357-454e-afd5-7bd5b42fb6ed\" (UID: \"ca06dcbb-f357-454e-afd5-7bd5b42fb6ed\") " Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.617339 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mcvs\" (UniqueName: \"kubernetes.io/projected/a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774-kube-api-access-4mcvs\") pod \"a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774\" (UID: \"a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774\") " Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.617434 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdfzk\" (UniqueName: \"kubernetes.io/projected/43f54580-b31f-45db-b046-3fb45235c384-kube-api-access-wdfzk\") pod \"43f54580-b31f-45db-b046-3fb45235c384\" (UID: \"43f54580-b31f-45db-b046-3fb45235c384\") " Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.617567 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmsld\" (UniqueName: \"kubernetes.io/projected/20f5475c-6d07-4be3-96f5-fed87f4ad62a-kube-api-access-dmsld\") pod \"20f5475c-6d07-4be3-96f5-fed87f4ad62a\" (UID: \"20f5475c-6d07-4be3-96f5-fed87f4ad62a\") " Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.617730 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774-operator-scripts\") pod \"a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774\" (UID: \"a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774\") " Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.617847 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfzbk\" (UniqueName: \"kubernetes.io/projected/37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3-kube-api-access-qfzbk\") pod \"37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3\" (UID: \"37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3\") " Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.617445 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20f5475c-6d07-4be3-96f5-fed87f4ad62a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "20f5475c-6d07-4be3-96f5-fed87f4ad62a" (UID: "20f5475c-6d07-4be3-96f5-fed87f4ad62a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.617837 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca06dcbb-f357-454e-afd5-7bd5b42fb6ed-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ca06dcbb-f357-454e-afd5-7bd5b42fb6ed" (UID: "ca06dcbb-f357-454e-afd5-7bd5b42fb6ed"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.618473 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cea49d65-9ca3-41a3-b970-25b0ccaac138-operator-scripts\") pod \"cea49d65-9ca3-41a3-b970-25b0ccaac138\" (UID: \"cea49d65-9ca3-41a3-b970-25b0ccaac138\") " Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.618652 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3-operator-scripts\") pod \"37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3\" (UID: \"37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3\") " Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.618731 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43f54580-b31f-45db-b046-3fb45235c384-operator-scripts\") pod \"43f54580-b31f-45db-b046-3fb45235c384\" (UID: \"43f54580-b31f-45db-b046-3fb45235c384\") " Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.618784 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrwqf\" (UniqueName: \"kubernetes.io/projected/ca06dcbb-f357-454e-afd5-7bd5b42fb6ed-kube-api-access-zrwqf\") pod \"ca06dcbb-f357-454e-afd5-7bd5b42fb6ed\" (UID: \"ca06dcbb-f357-454e-afd5-7bd5b42fb6ed\") " Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.618972 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774" (UID: "a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.619558 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3" (UID: "37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.620753 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca06dcbb-f357-454e-afd5-7bd5b42fb6ed-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.620856 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlhzg\" (UniqueName: \"kubernetes.io/projected/515a9b7f-6551-4a40-bd25-6c851129a6d0-kube-api-access-rlhzg\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.620869 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.620931 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/515a9b7f-6551-4a40-bd25-6c851129a6d0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.620941 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.620950 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20f5475c-6d07-4be3-96f5-fed87f4ad62a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.621295 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774-kube-api-access-4mcvs" (OuterVolumeSpecName: "kube-api-access-4mcvs") pod "a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774" (UID: "a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774"). InnerVolumeSpecName "kube-api-access-4mcvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.622921 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3-kube-api-access-qfzbk" (OuterVolumeSpecName: "kube-api-access-qfzbk") pod "37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3" (UID: "37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3"). InnerVolumeSpecName "kube-api-access-qfzbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.623306 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cea49d65-9ca3-41a3-b970-25b0ccaac138-kube-api-access-m4krs" (OuterVolumeSpecName: "kube-api-access-m4krs") pod "cea49d65-9ca3-41a3-b970-25b0ccaac138" (UID: "cea49d65-9ca3-41a3-b970-25b0ccaac138"). InnerVolumeSpecName "kube-api-access-m4krs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.623505 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43f54580-b31f-45db-b046-3fb45235c384-kube-api-access-wdfzk" (OuterVolumeSpecName: "kube-api-access-wdfzk") pod "43f54580-b31f-45db-b046-3fb45235c384" (UID: "43f54580-b31f-45db-b046-3fb45235c384"). InnerVolumeSpecName "kube-api-access-wdfzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.623747 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20f5475c-6d07-4be3-96f5-fed87f4ad62a-kube-api-access-dmsld" (OuterVolumeSpecName: "kube-api-access-dmsld") pod "20f5475c-6d07-4be3-96f5-fed87f4ad62a" (UID: "20f5475c-6d07-4be3-96f5-fed87f4ad62a"). InnerVolumeSpecName "kube-api-access-dmsld". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.623754 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43f54580-b31f-45db-b046-3fb45235c384-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "43f54580-b31f-45db-b046-3fb45235c384" (UID: "43f54580-b31f-45db-b046-3fb45235c384"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.624023 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca06dcbb-f357-454e-afd5-7bd5b42fb6ed-kube-api-access-zrwqf" (OuterVolumeSpecName: "kube-api-access-zrwqf") pod "ca06dcbb-f357-454e-afd5-7bd5b42fb6ed" (UID: "ca06dcbb-f357-454e-afd5-7bd5b42fb6ed"). InnerVolumeSpecName "kube-api-access-zrwqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.625101 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cea49d65-9ca3-41a3-b970-25b0ccaac138-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cea49d65-9ca3-41a3-b970-25b0ccaac138" (UID: "cea49d65-9ca3-41a3-b970-25b0ccaac138"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.723520 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfzbk\" (UniqueName: \"kubernetes.io/projected/37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3-kube-api-access-qfzbk\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.723551 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cea49d65-9ca3-41a3-b970-25b0ccaac138-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.723561 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43f54580-b31f-45db-b046-3fb45235c384-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.723569 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrwqf\" (UniqueName: \"kubernetes.io/projected/ca06dcbb-f357-454e-afd5-7bd5b42fb6ed-kube-api-access-zrwqf\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.723579 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4krs\" (UniqueName: \"kubernetes.io/projected/cea49d65-9ca3-41a3-b970-25b0ccaac138-kube-api-access-m4krs\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.723587 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mcvs\" (UniqueName: \"kubernetes.io/projected/a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774-kube-api-access-4mcvs\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.723596 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdfzk\" (UniqueName: \"kubernetes.io/projected/43f54580-b31f-45db-b046-3fb45235c384-kube-api-access-wdfzk\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.723604 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmsld\" (UniqueName: \"kubernetes.io/projected/20f5475c-6d07-4be3-96f5-fed87f4ad62a-kube-api-access-dmsld\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:23 crc kubenswrapper[4828]: I1210 19:19:23.801518 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2630e12-86b3-45d3-84c0-dd50f5cf8c15" path="/var/lib/kubelet/pods/d2630e12-86b3-45d3-84c0-dd50f5cf8c15/volumes" Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.068261 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zkd8f" event={"ID":"515a9b7f-6551-4a40-bd25-6c851129a6d0","Type":"ContainerDied","Data":"0aef5b1f7bee92f85c25b3b67be94bb1ac4ec9a5047706d4db8e7abed6a9bb6d"} Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.068547 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0aef5b1f7bee92f85c25b3b67be94bb1ac4ec9a5047706d4db8e7abed6a9bb6d" Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.068625 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zkd8f" Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.074514 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4btg4" event={"ID":"ca06dcbb-f357-454e-afd5-7bd5b42fb6ed","Type":"ContainerDied","Data":"0b062ed37673ad0899d9278be3fea7f9e350637a3e3480c5d1672a2b0a3b9a08"} Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.074556 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b062ed37673ad0899d9278be3fea7f9e350637a3e3480c5d1672a2b0a3b9a08" Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.074619 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4btg4" Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.079355 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-1a9d-account-create-update-ql2pg" event={"ID":"37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3","Type":"ContainerDied","Data":"067a00e41305df1c07cc77071effb7fd118e14897663ef4faeb897df9745f958"} Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.079390 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="067a00e41305df1c07cc77071effb7fd118e14897663ef4faeb897df9745f958" Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.079462 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-1a9d-account-create-update-ql2pg" Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.082886 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ae3c-account-create-update-v6g92" event={"ID":"43f54580-b31f-45db-b046-3fb45235c384","Type":"ContainerDied","Data":"0a735ceed02220bd65b1d7ffded2f84c12ff35a3bca815826bea3acae99f6ab8"} Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.082917 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a735ceed02220bd65b1d7ffded2f84c12ff35a3bca815826bea3acae99f6ab8" Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.082971 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ae3c-account-create-update-v6g92" Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.086526 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"5d6e68b4-1320-4111-a47d-51efcfab16b3","Type":"ContainerStarted","Data":"fa3df8df73f284d8f7201e4fec79d77e7bb356bf0a711980bb61f7279e4f91e8"} Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.089372 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-30e8-account-create-update-5zwls" Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.089447 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-30e8-account-create-update-5zwls" event={"ID":"20f5475c-6d07-4be3-96f5-fed87f4ad62a","Type":"ContainerDied","Data":"75cbada4d797e4865c6ff26b2f8810a88e582e6f12177fa6789a1d943d960bb8"} Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.089476 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75cbada4d797e4865c6ff26b2f8810a88e582e6f12177fa6789a1d943d960bb8" Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.092208 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-bbtgm" event={"ID":"a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774","Type":"ContainerDied","Data":"129f3009b77db2ef5f729d0825fd706f68d16e3e9e29d4ae0153599bb09a6065"} Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.092237 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="129f3009b77db2ef5f729d0825fd706f68d16e3e9e29d4ae0153599bb09a6065" Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.092291 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-bbtgm" Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.094650 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"9fd3d999-4b0d-422c-a7ae-01065d577c38","Type":"ContainerStarted","Data":"2a55dae265e67734536309d37d91f9df1cc16e0fbe174d434f83c5f5bda21d5f"} Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.096905 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d565-account-create-update-vwzdt" event={"ID":"cea49d65-9ca3-41a3-b970-25b0ccaac138","Type":"ContainerDied","Data":"37c0fd1c4a255f54a28b8d30226fe28b08bc7d14419a495ff9439f26a9b4b31e"} Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.096932 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37c0fd1c4a255f54a28b8d30226fe28b08bc7d14419a495ff9439f26a9b4b31e" Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.096988 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d565-account-create-update-vwzdt" Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.099439 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-b2rhf" event={"ID":"611d78ee-8203-4a4f-bb01-eedd92ab42fd","Type":"ContainerStarted","Data":"b7bf385ea061afcbcde12b0252ddb8b55001e1197495dfa1b2b45e8107778b13"} Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.102864 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e29bd910-2077-4ebe-8028-e810061fb704","Type":"ContainerStarted","Data":"3aefb32b262b96efe6e0a4c550398f1df82956c705933fd7aef62e26f71c198b"} Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.102896 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e29bd910-2077-4ebe-8028-e810061fb704","Type":"ContainerStarted","Data":"506c3134e1557b2088a0b2133965e886712a1491680bf33404d7372bef9a5987"} Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.102904 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e29bd910-2077-4ebe-8028-e810061fb704","Type":"ContainerStarted","Data":"37cb60a42a4248b89b4258bba1db2ef1e74a3701f80301c2271045f64c571224"} Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.127538 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=20.127522317 podStartE2EDuration="20.127522317s" podCreationTimestamp="2025-12-10 19:19:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:24.117485793 +0000 UTC m=+1444.628096798" watchObservedRunningTime="2025-12-10 19:19:24.127522317 +0000 UTC m=+1444.638133322" Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.147009 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-b2rhf" podStartSLOduration=5.494475907 podStartE2EDuration="10.14698926s" podCreationTimestamp="2025-12-10 19:19:14 +0000 UTC" firstStartedPulling="2025-12-10 19:19:18.544755666 +0000 UTC m=+1439.055366671" lastFinishedPulling="2025-12-10 19:19:23.197269019 +0000 UTC m=+1443.707880024" observedRunningTime="2025-12-10 19:19:24.135093307 +0000 UTC m=+1444.645704312" watchObservedRunningTime="2025-12-10 19:19:24.14698926 +0000 UTC m=+1444.657600265" Dec 10 19:19:24 crc kubenswrapper[4828]: I1210 19:19:24.174619 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=11.281454567 podStartE2EDuration="16.174450593s" podCreationTimestamp="2025-12-10 19:19:08 +0000 UTC" firstStartedPulling="2025-12-10 19:19:18.2627848 +0000 UTC m=+1438.773395805" lastFinishedPulling="2025-12-10 19:19:23.155780826 +0000 UTC m=+1443.666391831" observedRunningTime="2025-12-10 19:19:24.163683 +0000 UTC m=+1444.674294005" watchObservedRunningTime="2025-12-10 19:19:24.174450593 +0000 UTC m=+1444.685061598" Dec 10 19:19:25 crc kubenswrapper[4828]: I1210 19:19:25.113840 4828 generic.go:334] "Generic (PLEG): container finished" podID="5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2" containerID="5a669fd9dee5beaa44c3ac23322606f4e2ad87a45eca78cc6e7f974d677c7b6d" exitCode=0 Dec 10 19:19:25 crc kubenswrapper[4828]: I1210 19:19:25.114132 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9rgqj" event={"ID":"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2","Type":"ContainerDied","Data":"5a669fd9dee5beaa44c3ac23322606f4e2ad87a45eca78cc6e7f974d677c7b6d"} Dec 10 19:19:25 crc kubenswrapper[4828]: I1210 19:19:25.118667 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e29bd910-2077-4ebe-8028-e810061fb704","Type":"ContainerStarted","Data":"3e794e6c2a82ec59597aea9c2b4534587eb1c022180f182a333a2ea11b9c9d1f"} Dec 10 19:19:25 crc kubenswrapper[4828]: I1210 19:19:25.180459 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:25 crc kubenswrapper[4828]: E1210 19:19:25.190410 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f1645fa_ceff_4dcf_9b7e_e7b983a6ead2.slice/crio-5a669fd9dee5beaa44c3ac23322606f4e2ad87a45eca78cc6e7f974d677c7b6d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f1645fa_ceff_4dcf_9b7e_e7b983a6ead2.slice/crio-conmon-5a669fd9dee5beaa44c3ac23322606f4e2ad87a45eca78cc6e7f974d677c7b6d.scope\": RecentStats: unable to find data in memory cache]" Dec 10 19:19:26 crc kubenswrapper[4828]: I1210 19:19:26.129821 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e29bd910-2077-4ebe-8028-e810061fb704","Type":"ContainerStarted","Data":"bc7d2c59da2418858ba28f60a0d8fdcc46a2aa799785292efb3884650c694ce2"} Dec 10 19:19:26 crc kubenswrapper[4828]: I1210 19:19:26.628867 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9rgqj" Dec 10 19:19:26 crc kubenswrapper[4828]: I1210 19:19:26.786706 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-db-sync-config-data\") pod \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\" (UID: \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\") " Dec 10 19:19:26 crc kubenswrapper[4828]: I1210 19:19:26.786827 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-combined-ca-bundle\") pod \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\" (UID: \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\") " Dec 10 19:19:26 crc kubenswrapper[4828]: I1210 19:19:26.786927 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-config-data\") pod \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\" (UID: \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\") " Dec 10 19:19:26 crc kubenswrapper[4828]: I1210 19:19:26.787071 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bm6g\" (UniqueName: \"kubernetes.io/projected/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-kube-api-access-6bm6g\") pod \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\" (UID: \"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2\") " Dec 10 19:19:26 crc kubenswrapper[4828]: I1210 19:19:26.792485 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2" (UID: "5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:26 crc kubenswrapper[4828]: I1210 19:19:26.792627 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-kube-api-access-6bm6g" (OuterVolumeSpecName: "kube-api-access-6bm6g") pod "5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2" (UID: "5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2"). InnerVolumeSpecName "kube-api-access-6bm6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:26 crc kubenswrapper[4828]: I1210 19:19:26.814944 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2" (UID: "5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:26 crc kubenswrapper[4828]: I1210 19:19:26.837706 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-config-data" (OuterVolumeSpecName: "config-data") pod "5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2" (UID: "5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:26 crc kubenswrapper[4828]: I1210 19:19:26.889259 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bm6g\" (UniqueName: \"kubernetes.io/projected/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-kube-api-access-6bm6g\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:26 crc kubenswrapper[4828]: I1210 19:19:26.889290 4828 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:26 crc kubenswrapper[4828]: I1210 19:19:26.889302 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:26 crc kubenswrapper[4828]: I1210 19:19:26.889314 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.142483 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e29bd910-2077-4ebe-8028-e810061fb704","Type":"ContainerStarted","Data":"8aba21944b1167718d0721eb3909d32b4d627b845ef85bd6a4db3aa7856dcd0b"} Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.142855 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e29bd910-2077-4ebe-8028-e810061fb704","Type":"ContainerStarted","Data":"56c3147d55f1fc228fc030e1a074097a7dbbb73e9bf521077c66d24ed56a2eb8"} Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.142871 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e29bd910-2077-4ebe-8028-e810061fb704","Type":"ContainerStarted","Data":"f81f0a93ff42455b80b2b430b6a2dadb680f0f2a8c37b96238f46ad74569c8ad"} Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.145104 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-9rgqj" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.145163 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-9rgqj" event={"ID":"5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2","Type":"ContainerDied","Data":"a728cb42c785d2ddde237828275982bd43f39677e51e04e1f2a777c387be51c1"} Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.145230 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a728cb42c785d2ddde237828275982bd43f39677e51e04e1f2a777c387be51c1" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.151691 4828 generic.go:334] "Generic (PLEG): container finished" podID="611d78ee-8203-4a4f-bb01-eedd92ab42fd" containerID="b7bf385ea061afcbcde12b0252ddb8b55001e1197495dfa1b2b45e8107778b13" exitCode=0 Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.151746 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-b2rhf" event={"ID":"611d78ee-8203-4a4f-bb01-eedd92ab42fd","Type":"ContainerDied","Data":"b7bf385ea061afcbcde12b0252ddb8b55001e1197495dfa1b2b45e8107778b13"} Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.503027 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-6jtt2"] Dec 10 19:19:27 crc kubenswrapper[4828]: E1210 19:19:27.503549 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="515a9b7f-6551-4a40-bd25-6c851129a6d0" containerName="mariadb-database-create" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.503576 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="515a9b7f-6551-4a40-bd25-6c851129a6d0" containerName="mariadb-database-create" Dec 10 19:19:27 crc kubenswrapper[4828]: E1210 19:19:27.503587 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774" containerName="mariadb-database-create" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.503593 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774" containerName="mariadb-database-create" Dec 10 19:19:27 crc kubenswrapper[4828]: E1210 19:19:27.503604 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20f5475c-6d07-4be3-96f5-fed87f4ad62a" containerName="mariadb-account-create-update" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.503610 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="20f5475c-6d07-4be3-96f5-fed87f4ad62a" containerName="mariadb-account-create-update" Dec 10 19:19:27 crc kubenswrapper[4828]: E1210 19:19:27.503624 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2" containerName="glance-db-sync" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.503630 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2" containerName="glance-db-sync" Dec 10 19:19:27 crc kubenswrapper[4828]: E1210 19:19:27.503653 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43f54580-b31f-45db-b046-3fb45235c384" containerName="mariadb-account-create-update" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.503664 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="43f54580-b31f-45db-b046-3fb45235c384" containerName="mariadb-account-create-update" Dec 10 19:19:27 crc kubenswrapper[4828]: E1210 19:19:27.503678 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3" containerName="mariadb-account-create-update" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.503684 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3" containerName="mariadb-account-create-update" Dec 10 19:19:27 crc kubenswrapper[4828]: E1210 19:19:27.503693 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cea49d65-9ca3-41a3-b970-25b0ccaac138" containerName="mariadb-account-create-update" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.503699 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="cea49d65-9ca3-41a3-b970-25b0ccaac138" containerName="mariadb-account-create-update" Dec 10 19:19:27 crc kubenswrapper[4828]: E1210 19:19:27.503710 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="489d1351-3bec-4909-a5dd-7f5e14f280d6" containerName="mariadb-database-create" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.503716 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="489d1351-3bec-4909-a5dd-7f5e14f280d6" containerName="mariadb-database-create" Dec 10 19:19:27 crc kubenswrapper[4828]: E1210 19:19:27.503729 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca06dcbb-f357-454e-afd5-7bd5b42fb6ed" containerName="mariadb-database-create" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.503735 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca06dcbb-f357-454e-afd5-7bd5b42fb6ed" containerName="mariadb-database-create" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.503954 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="43f54580-b31f-45db-b046-3fb45235c384" containerName="mariadb-account-create-update" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.503973 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774" containerName="mariadb-database-create" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.503985 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3" containerName="mariadb-account-create-update" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.503993 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca06dcbb-f357-454e-afd5-7bd5b42fb6ed" containerName="mariadb-database-create" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.504012 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="cea49d65-9ca3-41a3-b970-25b0ccaac138" containerName="mariadb-account-create-update" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.504020 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="515a9b7f-6551-4a40-bd25-6c851129a6d0" containerName="mariadb-database-create" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.504031 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="489d1351-3bec-4909-a5dd-7f5e14f280d6" containerName="mariadb-database-create" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.504038 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2" containerName="glance-db-sync" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.504048 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="20f5475c-6d07-4be3-96f5-fed87f4ad62a" containerName="mariadb-account-create-update" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.505250 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.519448 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-6jtt2"] Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.603228 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-6jtt2\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.603318 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bz9b\" (UniqueName: \"kubernetes.io/projected/c664c515-ab2c-4211-b978-39bc88f62b03-kube-api-access-2bz9b\") pod \"dnsmasq-dns-74dc88fc-6jtt2\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.603422 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-dns-svc\") pod \"dnsmasq-dns-74dc88fc-6jtt2\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.603463 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-config\") pod \"dnsmasq-dns-74dc88fc-6jtt2\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.603489 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-6jtt2\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.705892 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bz9b\" (UniqueName: \"kubernetes.io/projected/c664c515-ab2c-4211-b978-39bc88f62b03-kube-api-access-2bz9b\") pod \"dnsmasq-dns-74dc88fc-6jtt2\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.706318 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-dns-svc\") pod \"dnsmasq-dns-74dc88fc-6jtt2\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.706357 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-config\") pod \"dnsmasq-dns-74dc88fc-6jtt2\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.706373 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-6jtt2\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.706511 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-6jtt2\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.707562 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-6jtt2\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.707572 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-dns-svc\") pod \"dnsmasq-dns-74dc88fc-6jtt2\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.707688 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-config\") pod \"dnsmasq-dns-74dc88fc-6jtt2\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.708131 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-6jtt2\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.737415 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bz9b\" (UniqueName: \"kubernetes.io/projected/c664c515-ab2c-4211-b978-39bc88f62b03-kube-api-access-2bz9b\") pod \"dnsmasq-dns-74dc88fc-6jtt2\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:27 crc kubenswrapper[4828]: I1210 19:19:27.838481 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:28 crc kubenswrapper[4828]: I1210 19:19:28.456094 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-6jtt2"] Dec 10 19:19:28 crc kubenswrapper[4828]: W1210 19:19:28.462411 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc664c515_ab2c_4211_b978_39bc88f62b03.slice/crio-876eb7223359878283ff72cf400015ac12821818e351e2980d9fd4bfe4c709f9 WatchSource:0}: Error finding container 876eb7223359878283ff72cf400015ac12821818e351e2980d9fd4bfe4c709f9: Status 404 returned error can't find the container with id 876eb7223359878283ff72cf400015ac12821818e351e2980d9fd4bfe4c709f9 Dec 10 19:19:28 crc kubenswrapper[4828]: I1210 19:19:28.572633 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-b2rhf" Dec 10 19:19:28 crc kubenswrapper[4828]: I1210 19:19:28.735636 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611d78ee-8203-4a4f-bb01-eedd92ab42fd-combined-ca-bundle\") pod \"611d78ee-8203-4a4f-bb01-eedd92ab42fd\" (UID: \"611d78ee-8203-4a4f-bb01-eedd92ab42fd\") " Dec 10 19:19:28 crc kubenswrapper[4828]: I1210 19:19:28.735782 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nv9q7\" (UniqueName: \"kubernetes.io/projected/611d78ee-8203-4a4f-bb01-eedd92ab42fd-kube-api-access-nv9q7\") pod \"611d78ee-8203-4a4f-bb01-eedd92ab42fd\" (UID: \"611d78ee-8203-4a4f-bb01-eedd92ab42fd\") " Dec 10 19:19:28 crc kubenswrapper[4828]: I1210 19:19:28.735867 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/611d78ee-8203-4a4f-bb01-eedd92ab42fd-config-data\") pod \"611d78ee-8203-4a4f-bb01-eedd92ab42fd\" (UID: \"611d78ee-8203-4a4f-bb01-eedd92ab42fd\") " Dec 10 19:19:28 crc kubenswrapper[4828]: I1210 19:19:28.762770 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/611d78ee-8203-4a4f-bb01-eedd92ab42fd-kube-api-access-nv9q7" (OuterVolumeSpecName: "kube-api-access-nv9q7") pod "611d78ee-8203-4a4f-bb01-eedd92ab42fd" (UID: "611d78ee-8203-4a4f-bb01-eedd92ab42fd"). InnerVolumeSpecName "kube-api-access-nv9q7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:28 crc kubenswrapper[4828]: I1210 19:19:28.838152 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nv9q7\" (UniqueName: \"kubernetes.io/projected/611d78ee-8203-4a4f-bb01-eedd92ab42fd-kube-api-access-nv9q7\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:28 crc kubenswrapper[4828]: I1210 19:19:28.896184 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/611d78ee-8203-4a4f-bb01-eedd92ab42fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "611d78ee-8203-4a4f-bb01-eedd92ab42fd" (UID: "611d78ee-8203-4a4f-bb01-eedd92ab42fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:28 crc kubenswrapper[4828]: I1210 19:19:28.938733 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/611d78ee-8203-4a4f-bb01-eedd92ab42fd-config-data" (OuterVolumeSpecName: "config-data") pod "611d78ee-8203-4a4f-bb01-eedd92ab42fd" (UID: "611d78ee-8203-4a4f-bb01-eedd92ab42fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:19:28 crc kubenswrapper[4828]: I1210 19:19:28.940356 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/611d78ee-8203-4a4f-bb01-eedd92ab42fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:28 crc kubenswrapper[4828]: I1210 19:19:28.940391 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/611d78ee-8203-4a4f-bb01-eedd92ab42fd-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.180114 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e29bd910-2077-4ebe-8028-e810061fb704","Type":"ContainerStarted","Data":"1bf92b84f8e10bb174afa2f66b5c4d3e64b3a3999a27572927c7ce9c8ed8250a"} Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.181458 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" event={"ID":"c664c515-ab2c-4211-b978-39bc88f62b03","Type":"ContainerStarted","Data":"876eb7223359878283ff72cf400015ac12821818e351e2980d9fd4bfe4c709f9"} Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.182948 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-b2rhf" event={"ID":"611d78ee-8203-4a4f-bb01-eedd92ab42fd","Type":"ContainerDied","Data":"cb2e663ddf153f81089beb1d3e599338f9f050b824d4df01e4756a19cf4344bd"} Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.182974 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb2e663ddf153f81089beb1d3e599338f9f050b824d4df01e4756a19cf4344bd" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.183062 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-b2rhf" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.390433 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-6jtt2"] Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.441939 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-d6bxn"] Dec 10 19:19:29 crc kubenswrapper[4828]: E1210 19:19:29.442855 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="611d78ee-8203-4a4f-bb01-eedd92ab42fd" containerName="keystone-db-sync" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.442994 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="611d78ee-8203-4a4f-bb01-eedd92ab42fd" containerName="keystone-db-sync" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.443342 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="611d78ee-8203-4a4f-bb01-eedd92ab42fd" containerName="keystone-db-sync" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.444312 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.453237 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.453714 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.453860 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-c922f" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.454030 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.464712 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.471791 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-d6bxn"] Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.487217 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d5679f497-qcgpq"] Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.494632 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.503307 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d5679f497-qcgpq"] Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.552576 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-fernet-keys\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.552671 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-combined-ca-bundle\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.552723 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-scripts\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.552751 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnghq\" (UniqueName: \"kubernetes.io/projected/6d698b70-20fe-4ee9-a919-56016dbfbc33-kube-api-access-xnghq\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.552782 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-credential-keys\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.552914 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-config-data\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.655131 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-dns-svc\") pod \"dnsmasq-dns-7d5679f497-qcgpq\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.655194 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-fernet-keys\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.655241 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-config\") pod \"dnsmasq-dns-7d5679f497-qcgpq\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.655264 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-combined-ca-bundle\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.655474 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-scripts\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.655556 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnghq\" (UniqueName: \"kubernetes.io/projected/6d698b70-20fe-4ee9-a919-56016dbfbc33-kube-api-access-xnghq\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.655596 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-ovsdbserver-nb\") pod \"dnsmasq-dns-7d5679f497-qcgpq\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.655634 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-credential-keys\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.655693 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-config-data\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.655720 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-ovsdbserver-sb\") pod \"dnsmasq-dns-7d5679f497-qcgpq\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.655783 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t9rw\" (UniqueName: \"kubernetes.io/projected/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-kube-api-access-6t9rw\") pod \"dnsmasq-dns-7d5679f497-qcgpq\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.659152 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-combined-ca-bundle\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.667179 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-scripts\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.674362 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-fernet-keys\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.677272 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-credential-keys\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.685530 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-config-data\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.702224 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnghq\" (UniqueName: \"kubernetes.io/projected/6d698b70-20fe-4ee9-a919-56016dbfbc33-kube-api-access-xnghq\") pod \"keystone-bootstrap-d6bxn\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.725865 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-2n7h2"] Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.727667 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2n7h2" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.741511 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.742154 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-jrdmz" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.758043 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-dns-svc\") pod \"dnsmasq-dns-7d5679f497-qcgpq\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.758150 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-config\") pod \"dnsmasq-dns-7d5679f497-qcgpq\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.758226 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-ovsdbserver-nb\") pod \"dnsmasq-dns-7d5679f497-qcgpq\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.758264 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-ovsdbserver-sb\") pod \"dnsmasq-dns-7d5679f497-qcgpq\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.758296 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t9rw\" (UniqueName: \"kubernetes.io/projected/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-kube-api-access-6t9rw\") pod \"dnsmasq-dns-7d5679f497-qcgpq\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.759409 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-config\") pod \"dnsmasq-dns-7d5679f497-qcgpq\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.759456 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-ovsdbserver-sb\") pod \"dnsmasq-dns-7d5679f497-qcgpq\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.759713 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-ovsdbserver-nb\") pod \"dnsmasq-dns-7d5679f497-qcgpq\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.759995 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-dns-svc\") pod \"dnsmasq-dns-7d5679f497-qcgpq\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.772969 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-n7jdh"] Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.775537 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.777011 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-n7jdh" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.777436 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-2n7h2"] Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.790390 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.790921 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.791046 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-n7jdh"] Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.791206 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-2s4gx" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.844767 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t9rw\" (UniqueName: \"kubernetes.io/projected/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-kube-api-access-6t9rw\") pod \"dnsmasq-dns-7d5679f497-qcgpq\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.870216 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acbabd11-f05a-4d18-a19a-6f1a74212a65-config-data\") pod \"heat-db-sync-2n7h2\" (UID: \"acbabd11-f05a-4d18-a19a-6f1a74212a65\") " pod="openstack/heat-db-sync-2n7h2" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.870413 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a56b6762-0c16-4b35-a11b-16e34810aeab-combined-ca-bundle\") pod \"neutron-db-sync-n7jdh\" (UID: \"a56b6762-0c16-4b35-a11b-16e34810aeab\") " pod="openstack/neutron-db-sync-n7jdh" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.870493 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a56b6762-0c16-4b35-a11b-16e34810aeab-config\") pod \"neutron-db-sync-n7jdh\" (UID: \"a56b6762-0c16-4b35-a11b-16e34810aeab\") " pod="openstack/neutron-db-sync-n7jdh" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.870544 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acbabd11-f05a-4d18-a19a-6f1a74212a65-combined-ca-bundle\") pod \"heat-db-sync-2n7h2\" (UID: \"acbabd11-f05a-4d18-a19a-6f1a74212a65\") " pod="openstack/heat-db-sync-2n7h2" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.870703 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrz7q\" (UniqueName: \"kubernetes.io/projected/acbabd11-f05a-4d18-a19a-6f1a74212a65-kube-api-access-qrz7q\") pod \"heat-db-sync-2n7h2\" (UID: \"acbabd11-f05a-4d18-a19a-6f1a74212a65\") " pod="openstack/heat-db-sync-2n7h2" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.870737 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l6qg\" (UniqueName: \"kubernetes.io/projected/a56b6762-0c16-4b35-a11b-16e34810aeab-kube-api-access-5l6qg\") pod \"neutron-db-sync-n7jdh\" (UID: \"a56b6762-0c16-4b35-a11b-16e34810aeab\") " pod="openstack/neutron-db-sync-n7jdh" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.938921 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d5679f497-qcgpq"] Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.940018 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.951346 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-nkxxv"] Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.970274 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nkxxv"] Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.970377 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nkxxv" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.974015 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acbabd11-f05a-4d18-a19a-6f1a74212a65-config-data\") pod \"heat-db-sync-2n7h2\" (UID: \"acbabd11-f05a-4d18-a19a-6f1a74212a65\") " pod="openstack/heat-db-sync-2n7h2" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.974087 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a56b6762-0c16-4b35-a11b-16e34810aeab-combined-ca-bundle\") pod \"neutron-db-sync-n7jdh\" (UID: \"a56b6762-0c16-4b35-a11b-16e34810aeab\") " pod="openstack/neutron-db-sync-n7jdh" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.974124 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a56b6762-0c16-4b35-a11b-16e34810aeab-config\") pod \"neutron-db-sync-n7jdh\" (UID: \"a56b6762-0c16-4b35-a11b-16e34810aeab\") " pod="openstack/neutron-db-sync-n7jdh" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.974152 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acbabd11-f05a-4d18-a19a-6f1a74212a65-combined-ca-bundle\") pod \"heat-db-sync-2n7h2\" (UID: \"acbabd11-f05a-4d18-a19a-6f1a74212a65\") " pod="openstack/heat-db-sync-2n7h2" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.974209 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrz7q\" (UniqueName: \"kubernetes.io/projected/acbabd11-f05a-4d18-a19a-6f1a74212a65-kube-api-access-qrz7q\") pod \"heat-db-sync-2n7h2\" (UID: \"acbabd11-f05a-4d18-a19a-6f1a74212a65\") " pod="openstack/heat-db-sync-2n7h2" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.974230 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l6qg\" (UniqueName: \"kubernetes.io/projected/a56b6762-0c16-4b35-a11b-16e34810aeab-kube-api-access-5l6qg\") pod \"neutron-db-sync-n7jdh\" (UID: \"a56b6762-0c16-4b35-a11b-16e34810aeab\") " pod="openstack/neutron-db-sync-n7jdh" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.977326 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.977463 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-8gm2q" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.977682 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.986127 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acbabd11-f05a-4d18-a19a-6f1a74212a65-combined-ca-bundle\") pod \"heat-db-sync-2n7h2\" (UID: \"acbabd11-f05a-4d18-a19a-6f1a74212a65\") " pod="openstack/heat-db-sync-2n7h2" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.986704 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a56b6762-0c16-4b35-a11b-16e34810aeab-config\") pod \"neutron-db-sync-n7jdh\" (UID: \"a56b6762-0c16-4b35-a11b-16e34810aeab\") " pod="openstack/neutron-db-sync-n7jdh" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.987063 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a56b6762-0c16-4b35-a11b-16e34810aeab-combined-ca-bundle\") pod \"neutron-db-sync-n7jdh\" (UID: \"a56b6762-0c16-4b35-a11b-16e34810aeab\") " pod="openstack/neutron-db-sync-n7jdh" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.993520 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56798b757f-df8js"] Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.995227 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:29 crc kubenswrapper[4828]: I1210 19:19:29.995656 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acbabd11-f05a-4d18-a19a-6f1a74212a65-config-data\") pod \"heat-db-sync-2n7h2\" (UID: \"acbabd11-f05a-4d18-a19a-6f1a74212a65\") " pod="openstack/heat-db-sync-2n7h2" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.012896 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56798b757f-df8js"] Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.025355 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrz7q\" (UniqueName: \"kubernetes.io/projected/acbabd11-f05a-4d18-a19a-6f1a74212a65-kube-api-access-qrz7q\") pod \"heat-db-sync-2n7h2\" (UID: \"acbabd11-f05a-4d18-a19a-6f1a74212a65\") " pod="openstack/heat-db-sync-2n7h2" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.043389 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l6qg\" (UniqueName: \"kubernetes.io/projected/a56b6762-0c16-4b35-a11b-16e34810aeab-kube-api-access-5l6qg\") pod \"neutron-db-sync-n7jdh\" (UID: \"a56b6762-0c16-4b35-a11b-16e34810aeab\") " pod="openstack/neutron-db-sync-n7jdh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.078228 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rskbl\" (UniqueName: \"kubernetes.io/projected/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-kube-api-access-rskbl\") pod \"placement-db-sync-nkxxv\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " pod="openstack/placement-db-sync-nkxxv" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.078278 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-combined-ca-bundle\") pod \"placement-db-sync-nkxxv\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " pod="openstack/placement-db-sync-nkxxv" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.078323 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-logs\") pod \"placement-db-sync-nkxxv\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " pod="openstack/placement-db-sync-nkxxv" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.078360 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-ovsdbserver-nb\") pod \"dnsmasq-dns-56798b757f-df8js\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.078425 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-config\") pod \"dnsmasq-dns-56798b757f-df8js\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.078453 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-scripts\") pod \"placement-db-sync-nkxxv\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " pod="openstack/placement-db-sync-nkxxv" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.078474 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-ovsdbserver-sb\") pod \"dnsmasq-dns-56798b757f-df8js\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.078505 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npkcf\" (UniqueName: \"kubernetes.io/projected/95e1cafb-8bc1-4448-8738-4be3d7596e72-kube-api-access-npkcf\") pod \"dnsmasq-dns-56798b757f-df8js\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.078520 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-config-data\") pod \"placement-db-sync-nkxxv\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " pod="openstack/placement-db-sync-nkxxv" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.078546 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-dns-svc\") pod \"dnsmasq-dns-56798b757f-df8js\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.126572 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2n7h2" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.196751 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-config\") pod \"dnsmasq-dns-56798b757f-df8js\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.196818 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-scripts\") pod \"placement-db-sync-nkxxv\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " pod="openstack/placement-db-sync-nkxxv" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.196838 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-ovsdbserver-sb\") pod \"dnsmasq-dns-56798b757f-df8js\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.196874 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npkcf\" (UniqueName: \"kubernetes.io/projected/95e1cafb-8bc1-4448-8738-4be3d7596e72-kube-api-access-npkcf\") pod \"dnsmasq-dns-56798b757f-df8js\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.196888 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-config-data\") pod \"placement-db-sync-nkxxv\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " pod="openstack/placement-db-sync-nkxxv" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.196915 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-dns-svc\") pod \"dnsmasq-dns-56798b757f-df8js\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.196975 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rskbl\" (UniqueName: \"kubernetes.io/projected/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-kube-api-access-rskbl\") pod \"placement-db-sync-nkxxv\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " pod="openstack/placement-db-sync-nkxxv" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.196995 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-combined-ca-bundle\") pod \"placement-db-sync-nkxxv\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " pod="openstack/placement-db-sync-nkxxv" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.197025 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-logs\") pod \"placement-db-sync-nkxxv\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " pod="openstack/placement-db-sync-nkxxv" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.197050 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-ovsdbserver-nb\") pod \"dnsmasq-dns-56798b757f-df8js\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.197922 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-ovsdbserver-nb\") pod \"dnsmasq-dns-56798b757f-df8js\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.198210 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-config\") pod \"dnsmasq-dns-56798b757f-df8js\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.198502 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-dns-svc\") pod \"dnsmasq-dns-56798b757f-df8js\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.198767 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-logs\") pod \"placement-db-sync-nkxxv\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " pod="openstack/placement-db-sync-nkxxv" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.199573 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-ovsdbserver-sb\") pod \"dnsmasq-dns-56798b757f-df8js\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.207167 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-scripts\") pod \"placement-db-sync-nkxxv\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " pod="openstack/placement-db-sync-nkxxv" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.223302 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-combined-ca-bundle\") pod \"placement-db-sync-nkxxv\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " pod="openstack/placement-db-sync-nkxxv" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.239511 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-config-data\") pod \"placement-db-sync-nkxxv\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " pod="openstack/placement-db-sync-nkxxv" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.244616 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rskbl\" (UniqueName: \"kubernetes.io/projected/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-kube-api-access-rskbl\") pod \"placement-db-sync-nkxxv\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " pod="openstack/placement-db-sync-nkxxv" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.252830 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npkcf\" (UniqueName: \"kubernetes.io/projected/95e1cafb-8bc1-4448-8738-4be3d7596e72-kube-api-access-npkcf\") pod \"dnsmasq-dns-56798b757f-df8js\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.265424 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-n7jdh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.287660 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nkxxv" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.330734 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.662751 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-2s89p"] Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.667512 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2s89p" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.670876 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.670972 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-gcmgv" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.697868 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-sqwjh"] Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.699348 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.705287 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.705531 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wbmsh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.705658 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.744302 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-2s89p"] Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.760097 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-sqwjh"] Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.771568 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-d6bxn"] Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.805191 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.807240 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.812097 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.812322 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.812498 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-mdn2k" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.826257 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.841196 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a634b8e-6ee2-4830-a764-2fcc5300b15d-combined-ca-bundle\") pod \"barbican-db-sync-2s89p\" (UID: \"1a634b8e-6ee2-4830-a764-2fcc5300b15d\") " pod="openstack/barbican-db-sync-2s89p" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.841283 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-db-sync-config-data\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.841331 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/22554fda-f55d-4243-8416-7362ae946396-etc-machine-id\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.841360 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1a634b8e-6ee2-4830-a764-2fcc5300b15d-db-sync-config-data\") pod \"barbican-db-sync-2s89p\" (UID: \"1a634b8e-6ee2-4830-a764-2fcc5300b15d\") " pod="openstack/barbican-db-sync-2s89p" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.841421 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-combined-ca-bundle\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.841458 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-config-data\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.841490 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-scripts\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.841530 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzt9p\" (UniqueName: \"kubernetes.io/projected/1a634b8e-6ee2-4830-a764-2fcc5300b15d-kube-api-access-qzt9p\") pod \"barbican-db-sync-2s89p\" (UID: \"1a634b8e-6ee2-4830-a764-2fcc5300b15d\") " pod="openstack/barbican-db-sync-2s89p" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.841560 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlh25\" (UniqueName: \"kubernetes.io/projected/22554fda-f55d-4243-8416-7362ae946396-kube-api-access-nlh25\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.869883 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.872464 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.875589 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.897580 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.943717 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8110268a-391c-479e-bb86-3bf7fdd0c987-logs\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.943769 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.943807 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-combined-ca-bundle\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.943974 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-config-data\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.944024 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-config-data\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.944065 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-scripts\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.944148 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzt9p\" (UniqueName: \"kubernetes.io/projected/1a634b8e-6ee2-4830-a764-2fcc5300b15d-kube-api-access-qzt9p\") pod \"barbican-db-sync-2s89p\" (UID: \"1a634b8e-6ee2-4830-a764-2fcc5300b15d\") " pod="openstack/barbican-db-sync-2s89p" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.944167 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-scripts\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.944203 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlh25\" (UniqueName: \"kubernetes.io/projected/22554fda-f55d-4243-8416-7362ae946396-kube-api-access-nlh25\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.944401 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a634b8e-6ee2-4830-a764-2fcc5300b15d-combined-ca-bundle\") pod \"barbican-db-sync-2s89p\" (UID: \"1a634b8e-6ee2-4830-a764-2fcc5300b15d\") " pod="openstack/barbican-db-sync-2s89p" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.944444 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8110268a-391c-479e-bb86-3bf7fdd0c987-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.944578 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-db-sync-config-data\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.944648 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/22554fda-f55d-4243-8416-7362ae946396-etc-machine-id\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.944697 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1a634b8e-6ee2-4830-a764-2fcc5300b15d-db-sync-config-data\") pod \"barbican-db-sync-2s89p\" (UID: \"1a634b8e-6ee2-4830-a764-2fcc5300b15d\") " pod="openstack/barbican-db-sync-2s89p" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.944742 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvwr4\" (UniqueName: \"kubernetes.io/projected/8110268a-391c-479e-bb86-3bf7fdd0c987-kube-api-access-qvwr4\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.944853 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.945658 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/22554fda-f55d-4243-8416-7362ae946396-etc-machine-id\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.955916 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1a634b8e-6ee2-4830-a764-2fcc5300b15d-db-sync-config-data\") pod \"barbican-db-sync-2s89p\" (UID: \"1a634b8e-6ee2-4830-a764-2fcc5300b15d\") " pod="openstack/barbican-db-sync-2s89p" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.963681 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a634b8e-6ee2-4830-a764-2fcc5300b15d-combined-ca-bundle\") pod \"barbican-db-sync-2s89p\" (UID: \"1a634b8e-6ee2-4830-a764-2fcc5300b15d\") " pod="openstack/barbican-db-sync-2s89p" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.966019 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlh25\" (UniqueName: \"kubernetes.io/projected/22554fda-f55d-4243-8416-7362ae946396-kube-api-access-nlh25\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.966945 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-config-data\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.967552 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-scripts\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.969471 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-combined-ca-bundle\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.972591 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-db-sync-config-data\") pod \"cinder-db-sync-sqwjh\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.972871 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-2n7h2"] Dec 10 19:19:30 crc kubenswrapper[4828]: I1210 19:19:30.980345 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzt9p\" (UniqueName: \"kubernetes.io/projected/1a634b8e-6ee2-4830-a764-2fcc5300b15d-kube-api-access-qzt9p\") pod \"barbican-db-sync-2s89p\" (UID: \"1a634b8e-6ee2-4830-a764-2fcc5300b15d\") " pod="openstack/barbican-db-sync-2s89p" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.047388 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-scripts\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.047535 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8110268a-391c-479e-bb86-3bf7fdd0c987-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.047572 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbpg2\" (UniqueName: \"kubernetes.io/projected/832260f7-4d4e-48ed-a72c-af9d73156d70-kube-api-access-hbpg2\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.047706 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-scripts\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.047774 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvwr4\" (UniqueName: \"kubernetes.io/projected/8110268a-391c-479e-bb86-3bf7fdd0c987-kube-api-access-qvwr4\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.047843 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-config-data\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.047897 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.047916 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/832260f7-4d4e-48ed-a72c-af9d73156d70-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.047966 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.047996 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8110268a-391c-479e-bb86-3bf7fdd0c987-logs\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.048016 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/832260f7-4d4e-48ed-a72c-af9d73156d70-logs\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.048038 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.048074 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-config-data\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.048098 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.048511 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8110268a-391c-479e-bb86-3bf7fdd0c987-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.048791 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8110268a-391c-479e-bb86-3bf7fdd0c987-logs\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.049113 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.049708 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2s89p" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.053033 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-scripts\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.056686 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-config-data\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.066789 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.073549 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.078347 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvwr4\" (UniqueName: \"kubernetes.io/projected/8110268a-391c-479e-bb86-3bf7fdd0c987-kube-api-access-qvwr4\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.118122 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " pod="openstack/glance-default-external-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.141104 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.152170 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.153325 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbpg2\" (UniqueName: \"kubernetes.io/projected/832260f7-4d4e-48ed-a72c-af9d73156d70-kube-api-access-hbpg2\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.153499 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-scripts\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.153623 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-config-data\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.153717 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.153825 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/832260f7-4d4e-48ed-a72c-af9d73156d70-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.153942 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/832260f7-4d4e-48ed-a72c-af9d73156d70-logs\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.154382 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.155129 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/832260f7-4d4e-48ed-a72c-af9d73156d70-logs\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.155329 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/832260f7-4d4e-48ed-a72c-af9d73156d70-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.164823 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.165360 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-scripts\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.170958 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-config-data\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.172024 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbpg2\" (UniqueName: \"kubernetes.io/projected/832260f7-4d4e-48ed-a72c-af9d73156d70-kube-api-access-hbpg2\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.209969 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.229744 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.309465 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.313316 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.319001 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.319327 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.329542 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.330915 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2n7h2" event={"ID":"acbabd11-f05a-4d18-a19a-6f1a74212a65","Type":"ContainerStarted","Data":"d21113b0f8bcddebee990f72dd664ff1bce1eba5fb5efd221d5984e4a358dc9f"} Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.354781 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-d6bxn" event={"ID":"6d698b70-20fe-4ee9-a919-56016dbfbc33","Type":"ContainerStarted","Data":"e74369545d7034e0cad0f4ac39e2d6b267ee6425d482c218662c4b4256709f2e"} Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.383293 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-nkxxv"] Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.420931 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56798b757f-df8js"] Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.433644 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-n7jdh"] Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.447219 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d5679f497-qcgpq"] Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.465851 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-scripts\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.465930 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.465962 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-config-data\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.465990 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cdea527-5341-46d2-b11f-9c88c4b35469-run-httpd\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.466064 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cdea527-5341-46d2-b11f-9c88c4b35469-log-httpd\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.466145 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.466225 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dntxb\" (UniqueName: \"kubernetes.io/projected/9cdea527-5341-46d2-b11f-9c88c4b35469-kube-api-access-dntxb\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: W1210 19:19:31.507211 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda56b6762_0c16_4b35_a11b_16e34810aeab.slice/crio-13654f6e07ca21e622fe41f927a38612118ee5833f606cacd4282906ae4cb5be WatchSource:0}: Error finding container 13654f6e07ca21e622fe41f927a38612118ee5833f606cacd4282906ae4cb5be: Status 404 returned error can't find the container with id 13654f6e07ca21e622fe41f927a38612118ee5833f606cacd4282906ae4cb5be Dec 10 19:19:31 crc kubenswrapper[4828]: W1210 19:19:31.510631 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4ef8a05_e672_4812_9f9a_f2ff51acf58e.slice/crio-af4e98facfc3757affbf6e6e1bc7b22b88970a712f53520626f62b00b1e4bc3f WatchSource:0}: Error finding container af4e98facfc3757affbf6e6e1bc7b22b88970a712f53520626f62b00b1e4bc3f: Status 404 returned error can't find the container with id af4e98facfc3757affbf6e6e1bc7b22b88970a712f53520626f62b00b1e4bc3f Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.567818 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.567889 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dntxb\" (UniqueName: \"kubernetes.io/projected/9cdea527-5341-46d2-b11f-9c88c4b35469-kube-api-access-dntxb\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.567981 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-scripts\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.568014 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.568032 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-config-data\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.568049 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cdea527-5341-46d2-b11f-9c88c4b35469-run-httpd\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.568095 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cdea527-5341-46d2-b11f-9c88c4b35469-log-httpd\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.568540 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cdea527-5341-46d2-b11f-9c88c4b35469-log-httpd\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.575697 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cdea527-5341-46d2-b11f-9c88c4b35469-run-httpd\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.578580 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.579410 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-scripts\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.582045 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.585844 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-config-data\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.600840 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dntxb\" (UniqueName: \"kubernetes.io/projected/9cdea527-5341-46d2-b11f-9c88c4b35469-kube-api-access-dntxb\") pod \"ceilometer-0\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.830168 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:19:31 crc kubenswrapper[4828]: W1210 19:19:31.902076 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a634b8e_6ee2_4830_a764_2fcc5300b15d.slice/crio-3add87f7232f1c5abc63d401be94d5333934f93b421e756db28435328e6f711d WatchSource:0}: Error finding container 3add87f7232f1c5abc63d401be94d5333934f93b421e756db28435328e6f711d: Status 404 returned error can't find the container with id 3add87f7232f1c5abc63d401be94d5333934f93b421e756db28435328e6f711d Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.916957 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-sqwjh"] Dec 10 19:19:31 crc kubenswrapper[4828]: I1210 19:19:31.929521 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-2s89p"] Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.248064 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.339234 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.389565 4828 generic.go:334] "Generic (PLEG): container finished" podID="95e1cafb-8bc1-4448-8738-4be3d7596e72" containerID="00e7c5f0ecf6fb8966aeaf550150cb81d215d281884ca137761f9b32e0555d99" exitCode=0 Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.390118 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56798b757f-df8js" event={"ID":"95e1cafb-8bc1-4448-8738-4be3d7596e72","Type":"ContainerDied","Data":"00e7c5f0ecf6fb8966aeaf550150cb81d215d281884ca137761f9b32e0555d99"} Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.390169 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56798b757f-df8js" event={"ID":"95e1cafb-8bc1-4448-8738-4be3d7596e72","Type":"ContainerStarted","Data":"2f16b3e08f526cf70a31cfec8160f100ddf4e5d63aadb36c081ce7216e2bd04a"} Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.414867 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-n7jdh" event={"ID":"a56b6762-0c16-4b35-a11b-16e34810aeab","Type":"ContainerStarted","Data":"c5f3ce88ff6d90ede4552698a8c5ec775f1d143ab5228560c41b371330fc2ff1"} Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.414908 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-n7jdh" event={"ID":"a56b6762-0c16-4b35-a11b-16e34810aeab","Type":"ContainerStarted","Data":"13654f6e07ca21e622fe41f927a38612118ee5833f606cacd4282906ae4cb5be"} Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.447277 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e29bd910-2077-4ebe-8028-e810061fb704","Type":"ContainerStarted","Data":"31180e1b9f1a90bad3cf779fb4ee57ac3313ded790b814e4ac9d85205fcdb745"} Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.447613 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e29bd910-2077-4ebe-8028-e810061fb704","Type":"ContainerStarted","Data":"ec36e5f14a58dd5398c824365f4342c7d79ee1fc145bf4b1bb49456affd95514"} Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.466404 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-sqwjh" event={"ID":"22554fda-f55d-4243-8416-7362ae946396","Type":"ContainerStarted","Data":"2d382ab9f15f453a6e87939a81e2d1205c250eb7ce031cdd0be141bea4a0a61d"} Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.478345 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nkxxv" event={"ID":"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b","Type":"ContainerStarted","Data":"c62774aab708c49f06edb6335d7fa2827ffaad464683a2d878d9e086e13703b1"} Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.484141 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8110268a-391c-479e-bb86-3bf7fdd0c987","Type":"ContainerStarted","Data":"e7606cb9ed8f8217944bd6b88f3a2c73d0e392c650bbd86c69bfb3b2a9fafb6c"} Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.485348 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.494681 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-n7jdh" podStartSLOduration=3.494653756 podStartE2EDuration="3.494653756s" podCreationTimestamp="2025-12-10 19:19:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:32.466071704 +0000 UTC m=+1452.976682709" watchObservedRunningTime="2025-12-10 19:19:32.494653756 +0000 UTC m=+1453.005264761" Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.503053 4828 generic.go:334] "Generic (PLEG): container finished" podID="e4ef8a05-e672-4812-9f9a-f2ff51acf58e" containerID="bdfa8c9bd14c12a436a1f785b08e6a6dbb318eef64ce567b926767efd5168a01" exitCode=0 Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.503145 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" event={"ID":"e4ef8a05-e672-4812-9f9a-f2ff51acf58e","Type":"ContainerDied","Data":"bdfa8c9bd14c12a436a1f785b08e6a6dbb318eef64ce567b926767efd5168a01"} Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.503172 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" event={"ID":"e4ef8a05-e672-4812-9f9a-f2ff51acf58e","Type":"ContainerStarted","Data":"af4e98facfc3757affbf6e6e1bc7b22b88970a712f53520626f62b00b1e4bc3f"} Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.509545 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-d6bxn" event={"ID":"6d698b70-20fe-4ee9-a919-56016dbfbc33","Type":"ContainerStarted","Data":"c1434316ba2de446c6bfb3c8f17f6eb29e3aa2954ec30915cc61ea0b8a9f2181"} Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.514159 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2s89p" event={"ID":"1a634b8e-6ee2-4830-a764-2fcc5300b15d","Type":"ContainerStarted","Data":"3add87f7232f1c5abc63d401be94d5333934f93b421e756db28435328e6f711d"} Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.523760 4828 generic.go:334] "Generic (PLEG): container finished" podID="c664c515-ab2c-4211-b978-39bc88f62b03" containerID="bc3e269e1c10282ecd7c93dd7be56aebe0574ee3190d2354929b83d3b46262b3" exitCode=0 Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.523828 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" event={"ID":"c664c515-ab2c-4211-b978-39bc88f62b03","Type":"ContainerDied","Data":"bc3e269e1c10282ecd7c93dd7be56aebe0574ee3190d2354929b83d3b46262b3"} Dec 10 19:19:32 crc kubenswrapper[4828]: I1210 19:19:32.552319 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-d6bxn" podStartSLOduration=3.552300035 podStartE2EDuration="3.552300035s" podCreationTimestamp="2025-12-10 19:19:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:32.547284342 +0000 UTC m=+1453.057895367" watchObservedRunningTime="2025-12-10 19:19:32.552300035 +0000 UTC m=+1453.062911040" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.259945 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.342260 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6t9rw\" (UniqueName: \"kubernetes.io/projected/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-kube-api-access-6t9rw\") pod \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.342394 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-config\") pod \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.342435 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-dns-svc\") pod \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.342599 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-ovsdbserver-nb\") pod \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.342651 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-ovsdbserver-sb\") pod \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\" (UID: \"e4ef8a05-e672-4812-9f9a-f2ff51acf58e\") " Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.357242 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-kube-api-access-6t9rw" (OuterVolumeSpecName: "kube-api-access-6t9rw") pod "e4ef8a05-e672-4812-9f9a-f2ff51acf58e" (UID: "e4ef8a05-e672-4812-9f9a-f2ff51acf58e"). InnerVolumeSpecName "kube-api-access-6t9rw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.378929 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-config" (OuterVolumeSpecName: "config") pod "e4ef8a05-e672-4812-9f9a-f2ff51acf58e" (UID: "e4ef8a05-e672-4812-9f9a-f2ff51acf58e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.399555 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e4ef8a05-e672-4812-9f9a-f2ff51acf58e" (UID: "e4ef8a05-e672-4812-9f9a-f2ff51acf58e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.403346 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e4ef8a05-e672-4812-9f9a-f2ff51acf58e" (UID: "e4ef8a05-e672-4812-9f9a-f2ff51acf58e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.405210 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.419375 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e4ef8a05-e672-4812-9f9a-f2ff51acf58e" (UID: "e4ef8a05-e672-4812-9f9a-f2ff51acf58e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.444814 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-dns-svc\") pod \"c664c515-ab2c-4211-b978-39bc88f62b03\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.444888 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-config\") pod \"c664c515-ab2c-4211-b978-39bc88f62b03\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.444972 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-ovsdbserver-sb\") pod \"c664c515-ab2c-4211-b978-39bc88f62b03\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.445059 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-ovsdbserver-nb\") pod \"c664c515-ab2c-4211-b978-39bc88f62b03\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.445190 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bz9b\" (UniqueName: \"kubernetes.io/projected/c664c515-ab2c-4211-b978-39bc88f62b03-kube-api-access-2bz9b\") pod \"c664c515-ab2c-4211-b978-39bc88f62b03\" (UID: \"c664c515-ab2c-4211-b978-39bc88f62b03\") " Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.445646 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6t9rw\" (UniqueName: \"kubernetes.io/projected/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-kube-api-access-6t9rw\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.445663 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.445674 4828 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.445685 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.445692 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e4ef8a05-e672-4812-9f9a-f2ff51acf58e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.449187 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c664c515-ab2c-4211-b978-39bc88f62b03-kube-api-access-2bz9b" (OuterVolumeSpecName: "kube-api-access-2bz9b") pod "c664c515-ab2c-4211-b978-39bc88f62b03" (UID: "c664c515-ab2c-4211-b978-39bc88f62b03"). InnerVolumeSpecName "kube-api-access-2bz9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.493688 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-config" (OuterVolumeSpecName: "config") pod "c664c515-ab2c-4211-b978-39bc88f62b03" (UID: "c664c515-ab2c-4211-b978-39bc88f62b03"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.540411 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.574373 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"832260f7-4d4e-48ed-a72c-af9d73156d70","Type":"ContainerStarted","Data":"82192848e612c9d7dd68570ea2e89696c016753b7cc32c1b59e46bf336b84ebd"} Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.576280 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" event={"ID":"c664c515-ab2c-4211-b978-39bc88f62b03","Type":"ContainerDied","Data":"876eb7223359878283ff72cf400015ac12821818e351e2980d9fd4bfe4c709f9"} Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.576336 4828 scope.go:117] "RemoveContainer" containerID="bc3e269e1c10282ecd7c93dd7be56aebe0574ee3190d2354929b83d3b46262b3" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.576555 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-6jtt2" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.586941 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.586986 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bz9b\" (UniqueName: \"kubernetes.io/projected/c664c515-ab2c-4211-b978-39bc88f62b03-kube-api-access-2bz9b\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.610415 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cdea527-5341-46d2-b11f-9c88c4b35469","Type":"ContainerStarted","Data":"2beb37740c7ec969b0b1259c125767b59cabcff63fcdc5935204f3c283007fbd"} Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.634275 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" event={"ID":"e4ef8a05-e672-4812-9f9a-f2ff51acf58e","Type":"ContainerDied","Data":"af4e98facfc3757affbf6e6e1bc7b22b88970a712f53520626f62b00b1e4bc3f"} Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.634373 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5679f497-qcgpq" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.674046 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c664c515-ab2c-4211-b978-39bc88f62b03" (UID: "c664c515-ab2c-4211-b978-39bc88f62b03"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.682713 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.690040 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.693741 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56798b757f-df8js" event={"ID":"95e1cafb-8bc1-4448-8738-4be3d7596e72","Type":"ContainerStarted","Data":"de0625c431ca982a1b2ea712717546aa2c5ec199edbccddc32655bb808b04bda"} Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.738835 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c664c515-ab2c-4211-b978-39bc88f62b03" (UID: "c664c515-ab2c-4211-b978-39bc88f62b03"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.743198 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c664c515-ab2c-4211-b978-39bc88f62b03" (UID: "c664c515-ab2c-4211-b978-39bc88f62b03"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.753767 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e29bd910-2077-4ebe-8028-e810061fb704","Type":"ContainerStarted","Data":"83a9924b07ad5446b4339358da35cef355bfc77c41251de185ff4152601b1f09"} Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.782027 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d5679f497-qcgpq"] Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.791854 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.791895 4828 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c664c515-ab2c-4211-b978-39bc88f62b03-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.813368 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d5679f497-qcgpq"] Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.838502 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.969057 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-6jtt2"] Dec 10 19:19:33 crc kubenswrapper[4828]: I1210 19:19:33.998036 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-6jtt2"] Dec 10 19:19:34 crc kubenswrapper[4828]: I1210 19:19:34.765464 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:34 crc kubenswrapper[4828]: I1210 19:19:34.788074 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56798b757f-df8js" podStartSLOduration=5.788055123 podStartE2EDuration="5.788055123s" podCreationTimestamp="2025-12-10 19:19:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:19:34.782552678 +0000 UTC m=+1455.293163683" watchObservedRunningTime="2025-12-10 19:19:34.788055123 +0000 UTC m=+1455.298666128" Dec 10 19:19:35 crc kubenswrapper[4828]: I1210 19:19:35.180650 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:35 crc kubenswrapper[4828]: I1210 19:19:35.187869 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:35 crc kubenswrapper[4828]: I1210 19:19:35.803890 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c664c515-ab2c-4211-b978-39bc88f62b03" path="/var/lib/kubelet/pods/c664c515-ab2c-4211-b978-39bc88f62b03/volumes" Dec 10 19:19:35 crc kubenswrapper[4828]: I1210 19:19:35.804558 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4ef8a05-e672-4812-9f9a-f2ff51acf58e" path="/var/lib/kubelet/pods/e4ef8a05-e672-4812-9f9a-f2ff51acf58e/volumes" Dec 10 19:19:35 crc kubenswrapper[4828]: I1210 19:19:35.805299 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 10 19:19:36 crc kubenswrapper[4828]: I1210 19:19:36.282753 4828 scope.go:117] "RemoveContainer" containerID="bdfa8c9bd14c12a436a1f785b08e6a6dbb318eef64ce567b926767efd5168a01" Dec 10 19:19:37 crc kubenswrapper[4828]: I1210 19:19:37.829875 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e29bd910-2077-4ebe-8028-e810061fb704","Type":"ContainerStarted","Data":"fb20cd58da838e87fe4997b76482d5b97522708f5008519496fff3112b154996"} Dec 10 19:19:37 crc kubenswrapper[4828]: I1210 19:19:37.835917 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"832260f7-4d4e-48ed-a72c-af9d73156d70","Type":"ContainerStarted","Data":"0b860ef7ec03f62883697c0aeab5dbc5fef529e10b76525a4d1d7be119c1e306"} Dec 10 19:19:37 crc kubenswrapper[4828]: I1210 19:19:37.838723 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8110268a-391c-479e-bb86-3bf7fdd0c987","Type":"ContainerStarted","Data":"f7cf4cd4cf43f81e166d49840007a9912685de8a42e6f2611af2b924d4af86e3"} Dec 10 19:19:40 crc kubenswrapper[4828]: I1210 19:19:40.332954 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:19:40 crc kubenswrapper[4828]: I1210 19:19:40.416472 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-lh7pb"] Dec 10 19:19:40 crc kubenswrapper[4828]: I1210 19:19:40.417600 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" podUID="8e282513-5ba6-4eec-8738-8406c723a69b" containerName="dnsmasq-dns" containerID="cri-o://f4ec0d6ab2875bf85d211db7bb4df48fce108ac5720f693b4635f071c1bdb002" gracePeriod=10 Dec 10 19:19:40 crc kubenswrapper[4828]: I1210 19:19:40.875410 4828 generic.go:334] "Generic (PLEG): container finished" podID="8e282513-5ba6-4eec-8738-8406c723a69b" containerID="f4ec0d6ab2875bf85d211db7bb4df48fce108ac5720f693b4635f071c1bdb002" exitCode=0 Dec 10 19:19:40 crc kubenswrapper[4828]: I1210 19:19:40.875453 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" event={"ID":"8e282513-5ba6-4eec-8738-8406c723a69b","Type":"ContainerDied","Data":"f4ec0d6ab2875bf85d211db7bb4df48fce108ac5720f693b4635f071c1bdb002"} Dec 10 19:19:40 crc kubenswrapper[4828]: I1210 19:19:40.877025 4828 generic.go:334] "Generic (PLEG): container finished" podID="6d698b70-20fe-4ee9-a919-56016dbfbc33" containerID="c1434316ba2de446c6bfb3c8f17f6eb29e3aa2954ec30915cc61ea0b8a9f2181" exitCode=0 Dec 10 19:19:40 crc kubenswrapper[4828]: I1210 19:19:40.877063 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-d6bxn" event={"ID":"6d698b70-20fe-4ee9-a919-56016dbfbc33","Type":"ContainerDied","Data":"c1434316ba2de446c6bfb3c8f17f6eb29e3aa2954ec30915cc61ea0b8a9f2181"} Dec 10 19:19:43 crc kubenswrapper[4828]: I1210 19:19:43.872364 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" podUID="8e282513-5ba6-4eec-8738-8406c723a69b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.146:5353: connect: connection refused" Dec 10 19:19:48 crc kubenswrapper[4828]: I1210 19:19:48.871806 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" podUID="8e282513-5ba6-4eec-8738-8406c723a69b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.146:5353: connect: connection refused" Dec 10 19:19:51 crc kubenswrapper[4828]: I1210 19:19:51.230384 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:19:51 crc kubenswrapper[4828]: I1210 19:19:51.230834 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:19:53 crc kubenswrapper[4828]: I1210 19:19:53.870844 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" podUID="8e282513-5ba6-4eec-8738-8406c723a69b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.146:5353: connect: connection refused" Dec 10 19:19:53 crc kubenswrapper[4828]: I1210 19:19:53.871269 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:19:58 crc kubenswrapper[4828]: I1210 19:19:58.871510 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" podUID="8e282513-5ba6-4eec-8738-8406c723a69b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.146:5353: connect: connection refused" Dec 10 19:20:02 crc kubenswrapper[4828]: E1210 19:20:02.251912 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Dec 10 19:20:02 crc kubenswrapper[4828]: E1210 19:20:02.252560 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rskbl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-nkxxv_openstack(d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:20:02 crc kubenswrapper[4828]: E1210 19:20:02.253772 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-nkxxv" podUID="d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b" Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.343029 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.448502 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-credential-keys\") pod \"6d698b70-20fe-4ee9-a919-56016dbfbc33\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.448772 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-config-data\") pod \"6d698b70-20fe-4ee9-a919-56016dbfbc33\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.448855 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-combined-ca-bundle\") pod \"6d698b70-20fe-4ee9-a919-56016dbfbc33\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.448928 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-scripts\") pod \"6d698b70-20fe-4ee9-a919-56016dbfbc33\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.449018 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnghq\" (UniqueName: \"kubernetes.io/projected/6d698b70-20fe-4ee9-a919-56016dbfbc33-kube-api-access-xnghq\") pod \"6d698b70-20fe-4ee9-a919-56016dbfbc33\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.449171 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-fernet-keys\") pod \"6d698b70-20fe-4ee9-a919-56016dbfbc33\" (UID: \"6d698b70-20fe-4ee9-a919-56016dbfbc33\") " Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.454310 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6d698b70-20fe-4ee9-a919-56016dbfbc33" (UID: "6d698b70-20fe-4ee9-a919-56016dbfbc33"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.455128 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d698b70-20fe-4ee9-a919-56016dbfbc33-kube-api-access-xnghq" (OuterVolumeSpecName: "kube-api-access-xnghq") pod "6d698b70-20fe-4ee9-a919-56016dbfbc33" (UID: "6d698b70-20fe-4ee9-a919-56016dbfbc33"). InnerVolumeSpecName "kube-api-access-xnghq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.455307 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6d698b70-20fe-4ee9-a919-56016dbfbc33" (UID: "6d698b70-20fe-4ee9-a919-56016dbfbc33"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.457238 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-scripts" (OuterVolumeSpecName: "scripts") pod "6d698b70-20fe-4ee9-a919-56016dbfbc33" (UID: "6d698b70-20fe-4ee9-a919-56016dbfbc33"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.490490 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-config-data" (OuterVolumeSpecName: "config-data") pod "6d698b70-20fe-4ee9-a919-56016dbfbc33" (UID: "6d698b70-20fe-4ee9-a919-56016dbfbc33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.492039 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d698b70-20fe-4ee9-a919-56016dbfbc33" (UID: "6d698b70-20fe-4ee9-a919-56016dbfbc33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.551705 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.551742 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.551751 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnghq\" (UniqueName: \"kubernetes.io/projected/6d698b70-20fe-4ee9-a919-56016dbfbc33-kube-api-access-xnghq\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.551762 4828 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.551771 4828 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:02 crc kubenswrapper[4828]: I1210 19:20:02.551780 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6d698b70-20fe-4ee9-a919-56016dbfbc33-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:02 crc kubenswrapper[4828]: E1210 19:20:02.634187 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Dec 10 19:20:02 crc kubenswrapper[4828]: E1210 19:20:02.634539 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qrz7q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-2n7h2_openstack(acbabd11-f05a-4d18-a19a-6f1a74212a65): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:20:02 crc kubenswrapper[4828]: E1210 19:20:02.635735 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-2n7h2" podUID="acbabd11-f05a-4d18-a19a-6f1a74212a65" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.137861 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-d6bxn" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.138110 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-d6bxn" event={"ID":"6d698b70-20fe-4ee9-a919-56016dbfbc33","Type":"ContainerDied","Data":"e74369545d7034e0cad0f4ac39e2d6b267ee6425d482c218662c4b4256709f2e"} Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.138141 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e74369545d7034e0cad0f4ac39e2d6b267ee6425d482c218662c4b4256709f2e" Dec 10 19:20:03 crc kubenswrapper[4828]: E1210 19:20:03.138967 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-2n7h2" podUID="acbabd11-f05a-4d18-a19a-6f1a74212a65" Dec 10 19:20:03 crc kubenswrapper[4828]: E1210 19:20:03.141010 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-nkxxv" podUID="d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.441361 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-d6bxn"] Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.451881 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-d6bxn"] Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.548789 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-gwbxv"] Dec 10 19:20:03 crc kubenswrapper[4828]: E1210 19:20:03.549932 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c664c515-ab2c-4211-b978-39bc88f62b03" containerName="init" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.549954 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c664c515-ab2c-4211-b978-39bc88f62b03" containerName="init" Dec 10 19:20:03 crc kubenswrapper[4828]: E1210 19:20:03.549971 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d698b70-20fe-4ee9-a919-56016dbfbc33" containerName="keystone-bootstrap" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.549979 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d698b70-20fe-4ee9-a919-56016dbfbc33" containerName="keystone-bootstrap" Dec 10 19:20:03 crc kubenswrapper[4828]: E1210 19:20:03.550020 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4ef8a05-e672-4812-9f9a-f2ff51acf58e" containerName="init" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.550029 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4ef8a05-e672-4812-9f9a-f2ff51acf58e" containerName="init" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.550254 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c664c515-ab2c-4211-b978-39bc88f62b03" containerName="init" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.550287 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4ef8a05-e672-4812-9f9a-f2ff51acf58e" containerName="init" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.550300 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d698b70-20fe-4ee9-a919-56016dbfbc33" containerName="keystone-bootstrap" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.551104 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.553190 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.554315 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.554568 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.554690 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.554769 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-c922f" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.562196 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-gwbxv"] Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.577114 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-config-data\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.577248 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-combined-ca-bundle\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.577294 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-credential-keys\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.577582 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7h7k\" (UniqueName: \"kubernetes.io/projected/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-kube-api-access-t7h7k\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.577822 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-scripts\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.577854 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-fernet-keys\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.679661 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7h7k\" (UniqueName: \"kubernetes.io/projected/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-kube-api-access-t7h7k\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.680007 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-scripts\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.680116 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-fernet-keys\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.681679 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-config-data\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.681950 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-combined-ca-bundle\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.682112 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-credential-keys\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.698480 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-credential-keys\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.698535 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-fernet-keys\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.700536 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-scripts\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.701446 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-config-data\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.702041 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-combined-ca-bundle\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.704492 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7h7k\" (UniqueName: \"kubernetes.io/projected/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-kube-api-access-t7h7k\") pod \"keystone-bootstrap-gwbxv\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:03 crc kubenswrapper[4828]: E1210 19:20:03.767635 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:800aa67a8f7a6cd334bd8ba44d1410e8aebc3aa6f347c5abc168de1222420526: Get \"https://quay.io/v2/podified-antelope-centos9/openstack-ceilometer-central/blobs/sha256:800aa67a8f7a6cd334bd8ba44d1410e8aebc3aa6f347c5abc168de1222420526\": context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 10 19:20:03 crc kubenswrapper[4828]: E1210 19:20:03.767817 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n549h5dbh5f5h659h55bh645hcdh9dhbh5fbh78h594h5ffhcch5ffh5c8h696hbfh9dhbch579hfbhf9hbch549h89hfh667h685h598h7bh6fq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dntxb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9cdea527-5341-46d2-b11f-9c88c4b35469): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:800aa67a8f7a6cd334bd8ba44d1410e8aebc3aa6f347c5abc168de1222420526: Get \"https://quay.io/v2/podified-antelope-centos9/openstack-ceilometer-central/blobs/sha256:800aa67a8f7a6cd334bd8ba44d1410e8aebc3aa6f347c5abc168de1222420526\": context canceled" logger="UnhandledError" Dec 10 19:20:03 crc kubenswrapper[4828]: E1210 19:20:03.813474 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 10 19:20:03 crc kubenswrapper[4828]: E1210 19:20:03.813621 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nlh25,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-sqwjh_openstack(22554fda-f55d-4243-8416-7362ae946396): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:20:03 crc kubenswrapper[4828]: E1210 19:20:03.814817 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-sqwjh" podUID="22554fda-f55d-4243-8416-7362ae946396" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.815627 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d698b70-20fe-4ee9-a919-56016dbfbc33" path="/var/lib/kubelet/pods/6d698b70-20fe-4ee9-a919-56016dbfbc33/volumes" Dec 10 19:20:03 crc kubenswrapper[4828]: I1210 19:20:03.883329 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.014616 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.172772 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" event={"ID":"8e282513-5ba6-4eec-8738-8406c723a69b","Type":"ContainerDied","Data":"9f5d1d75fd05a15d87c591fed65746e6786b2f811e1de06682e3a922e0eb0571"} Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.173154 4828 scope.go:117] "RemoveContainer" containerID="f4ec0d6ab2875bf85d211db7bb4df48fce108ac5720f693b4635f071c1bdb002" Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.172854 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" Dec 10 19:20:04 crc kubenswrapper[4828]: E1210 19:20:04.175001 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-sqwjh" podUID="22554fda-f55d-4243-8416-7362ae946396" Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.193783 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wkzn\" (UniqueName: \"kubernetes.io/projected/8e282513-5ba6-4eec-8738-8406c723a69b-kube-api-access-6wkzn\") pod \"8e282513-5ba6-4eec-8738-8406c723a69b\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.193914 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-config\") pod \"8e282513-5ba6-4eec-8738-8406c723a69b\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.194027 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-dns-svc\") pod \"8e282513-5ba6-4eec-8738-8406c723a69b\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.194094 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-ovsdbserver-nb\") pod \"8e282513-5ba6-4eec-8738-8406c723a69b\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.194203 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-ovsdbserver-sb\") pod \"8e282513-5ba6-4eec-8738-8406c723a69b\" (UID: \"8e282513-5ba6-4eec-8738-8406c723a69b\") " Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.214121 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e282513-5ba6-4eec-8738-8406c723a69b-kube-api-access-6wkzn" (OuterVolumeSpecName: "kube-api-access-6wkzn") pod "8e282513-5ba6-4eec-8738-8406c723a69b" (UID: "8e282513-5ba6-4eec-8738-8406c723a69b"). InnerVolumeSpecName "kube-api-access-6wkzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.308004 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wkzn\" (UniqueName: \"kubernetes.io/projected/8e282513-5ba6-4eec-8738-8406c723a69b-kube-api-access-6wkzn\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.320011 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-config" (OuterVolumeSpecName: "config") pod "8e282513-5ba6-4eec-8738-8406c723a69b" (UID: "8e282513-5ba6-4eec-8738-8406c723a69b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.335681 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8e282513-5ba6-4eec-8738-8406c723a69b" (UID: "8e282513-5ba6-4eec-8738-8406c723a69b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.368819 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8e282513-5ba6-4eec-8738-8406c723a69b" (UID: "8e282513-5ba6-4eec-8738-8406c723a69b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.371775 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8e282513-5ba6-4eec-8738-8406c723a69b" (UID: "8e282513-5ba6-4eec-8738-8406c723a69b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.385320 4828 scope.go:117] "RemoveContainer" containerID="73fd9fac71d50343b5ae049985abafd0a4bbf2be45a1702f59cc2b82d9fd755a" Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.410781 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.410845 4828 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.410856 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.410867 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e282513-5ba6-4eec-8738-8406c723a69b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.520635 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-lh7pb"] Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.533748 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-lh7pb"] Dec 10 19:20:04 crc kubenswrapper[4828]: I1210 19:20:04.680722 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-gwbxv"] Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.187910 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"832260f7-4d4e-48ed-a72c-af9d73156d70","Type":"ContainerStarted","Data":"2034e9c324dd4b2994000f8e4b97e94ee61931119f5b59356d6f9a63990e46a1"} Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.188231 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="832260f7-4d4e-48ed-a72c-af9d73156d70" containerName="glance-log" containerID="cri-o://0b860ef7ec03f62883697c0aeab5dbc5fef529e10b76525a4d1d7be119c1e306" gracePeriod=30 Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.188609 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="832260f7-4d4e-48ed-a72c-af9d73156d70" containerName="glance-httpd" containerID="cri-o://2034e9c324dd4b2994000f8e4b97e94ee61931119f5b59356d6f9a63990e46a1" gracePeriod=30 Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.191303 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8110268a-391c-479e-bb86-3bf7fdd0c987","Type":"ContainerStarted","Data":"35a12ed90ec6d821670de8a0a219594bf0d6168162509a1b23e43b4bcfbbbd66"} Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.191385 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8110268a-391c-479e-bb86-3bf7fdd0c987" containerName="glance-log" containerID="cri-o://f7cf4cd4cf43f81e166d49840007a9912685de8a42e6f2611af2b924d4af86e3" gracePeriod=30 Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.191475 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8110268a-391c-479e-bb86-3bf7fdd0c987" containerName="glance-httpd" containerID="cri-o://35a12ed90ec6d821670de8a0a219594bf0d6168162509a1b23e43b4bcfbbbd66" gracePeriod=30 Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.195483 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gwbxv" event={"ID":"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a","Type":"ContainerStarted","Data":"fc6c64ed594fe07d98e5a50b61dc2fe5a667292c054740f5de9c8d1fe0244a74"} Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.195525 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gwbxv" event={"ID":"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a","Type":"ContainerStarted","Data":"5722fad71fd4a0a6de3283be7eb278f4e60e48a95598f6cb0867cb45060e76d6"} Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.198548 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2s89p" event={"ID":"1a634b8e-6ee2-4830-a764-2fcc5300b15d","Type":"ContainerStarted","Data":"da2f7b7fd47c67b352b34c029147284bbad04762c536da585a6f875dbf0617d8"} Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.215756 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e29bd910-2077-4ebe-8028-e810061fb704","Type":"ContainerStarted","Data":"19c99561fae4f91a04e46e2ce15b76bb129374dd4dffc6f8b1b43a37f2113024"} Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.215819 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e29bd910-2077-4ebe-8028-e810061fb704","Type":"ContainerStarted","Data":"0d100919578e05bbaeeebaf0a12066b60235029ae59fcc7e913e37fc5abcb0e5"} Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.216788 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=36.216766684 podStartE2EDuration="36.216766684s" podCreationTimestamp="2025-12-10 19:19:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:05.211850165 +0000 UTC m=+1485.722461170" watchObservedRunningTime="2025-12-10 19:20:05.216766684 +0000 UTC m=+1485.727377689" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.279257 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-gwbxv" podStartSLOduration=2.279229149 podStartE2EDuration="2.279229149s" podCreationTimestamp="2025-12-10 19:20:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:05.246067796 +0000 UTC m=+1485.756678811" watchObservedRunningTime="2025-12-10 19:20:05.279229149 +0000 UTC m=+1485.789840164" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.281144 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=36.281132759 podStartE2EDuration="36.281132759s" podCreationTimestamp="2025-12-10 19:19:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:05.268043335 +0000 UTC m=+1485.778654330" watchObservedRunningTime="2025-12-10 19:20:05.281132759 +0000 UTC m=+1485.791743764" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.297579 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-2s89p" podStartSLOduration=3.406618263 podStartE2EDuration="35.297560652s" podCreationTimestamp="2025-12-10 19:19:30 +0000 UTC" firstStartedPulling="2025-12-10 19:19:31.912110835 +0000 UTC m=+1452.422721830" lastFinishedPulling="2025-12-10 19:20:03.803053214 +0000 UTC m=+1484.313664219" observedRunningTime="2025-12-10 19:20:05.284286572 +0000 UTC m=+1485.794897587" watchObservedRunningTime="2025-12-10 19:20:05.297560652 +0000 UTC m=+1485.808171657" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.339280 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=77.964397102 podStartE2EDuration="1m27.339259339s" podCreationTimestamp="2025-12-10 19:18:38 +0000 UTC" firstStartedPulling="2025-12-10 19:19:18.629447706 +0000 UTC m=+1439.140058711" lastFinishedPulling="2025-12-10 19:19:28.004309943 +0000 UTC m=+1448.514920948" observedRunningTime="2025-12-10 19:20:05.331157186 +0000 UTC m=+1485.841768191" watchObservedRunningTime="2025-12-10 19:20:05.339259339 +0000 UTC m=+1485.849870344" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.616250 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-ptl7r"] Dec 10 19:20:05 crc kubenswrapper[4828]: E1210 19:20:05.617116 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e282513-5ba6-4eec-8738-8406c723a69b" containerName="dnsmasq-dns" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.617131 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e282513-5ba6-4eec-8738-8406c723a69b" containerName="dnsmasq-dns" Dec 10 19:20:05 crc kubenswrapper[4828]: E1210 19:20:05.617153 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e282513-5ba6-4eec-8738-8406c723a69b" containerName="init" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.617160 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e282513-5ba6-4eec-8738-8406c723a69b" containerName="init" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.617395 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e282513-5ba6-4eec-8738-8406c723a69b" containerName="dnsmasq-dns" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.618787 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.621595 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.653376 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-ptl7r"] Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.745328 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.745384 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.745457 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-config\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.745507 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq76g\" (UniqueName: \"kubernetes.io/projected/046c39f9-a9eb-45d4-9965-5575059fdcfc-kube-api-access-lq76g\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.745526 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.745569 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.804912 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e282513-5ba6-4eec-8738-8406c723a69b" path="/var/lib/kubelet/pods/8e282513-5ba6-4eec-8738-8406c723a69b/volumes" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.848124 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.849070 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.849207 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.849303 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-config\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.849418 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq76g\" (UniqueName: \"kubernetes.io/projected/046c39f9-a9eb-45d4-9965-5575059fdcfc-kube-api-access-lq76g\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.849450 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.849548 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.849940 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.850423 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.850839 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.852446 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-config\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.870411 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq76g\" (UniqueName: \"kubernetes.io/projected/046c39f9-a9eb-45d4-9965-5575059fdcfc-kube-api-access-lq76g\") pod \"dnsmasq-dns-56df8fb6b7-ptl7r\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:05 crc kubenswrapper[4828]: I1210 19:20:05.953083 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:06 crc kubenswrapper[4828]: I1210 19:20:06.230887 4828 generic.go:334] "Generic (PLEG): container finished" podID="8110268a-391c-479e-bb86-3bf7fdd0c987" containerID="35a12ed90ec6d821670de8a0a219594bf0d6168162509a1b23e43b4bcfbbbd66" exitCode=0 Dec 10 19:20:06 crc kubenswrapper[4828]: I1210 19:20:06.230938 4828 generic.go:334] "Generic (PLEG): container finished" podID="8110268a-391c-479e-bb86-3bf7fdd0c987" containerID="f7cf4cd4cf43f81e166d49840007a9912685de8a42e6f2611af2b924d4af86e3" exitCode=143 Dec 10 19:20:06 crc kubenswrapper[4828]: I1210 19:20:06.230977 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8110268a-391c-479e-bb86-3bf7fdd0c987","Type":"ContainerDied","Data":"35a12ed90ec6d821670de8a0a219594bf0d6168162509a1b23e43b4bcfbbbd66"} Dec 10 19:20:06 crc kubenswrapper[4828]: I1210 19:20:06.231042 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8110268a-391c-479e-bb86-3bf7fdd0c987","Type":"ContainerDied","Data":"f7cf4cd4cf43f81e166d49840007a9912685de8a42e6f2611af2b924d4af86e3"} Dec 10 19:20:06 crc kubenswrapper[4828]: I1210 19:20:06.234297 4828 generic.go:334] "Generic (PLEG): container finished" podID="832260f7-4d4e-48ed-a72c-af9d73156d70" containerID="2034e9c324dd4b2994000f8e4b97e94ee61931119f5b59356d6f9a63990e46a1" exitCode=0 Dec 10 19:20:06 crc kubenswrapper[4828]: I1210 19:20:06.234332 4828 generic.go:334] "Generic (PLEG): container finished" podID="832260f7-4d4e-48ed-a72c-af9d73156d70" containerID="0b860ef7ec03f62883697c0aeab5dbc5fef529e10b76525a4d1d7be119c1e306" exitCode=143 Dec 10 19:20:06 crc kubenswrapper[4828]: I1210 19:20:06.235465 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"832260f7-4d4e-48ed-a72c-af9d73156d70","Type":"ContainerDied","Data":"2034e9c324dd4b2994000f8e4b97e94ee61931119f5b59356d6f9a63990e46a1"} Dec 10 19:20:06 crc kubenswrapper[4828]: I1210 19:20:06.235501 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"832260f7-4d4e-48ed-a72c-af9d73156d70","Type":"ContainerDied","Data":"0b860ef7ec03f62883697c0aeab5dbc5fef529e10b76525a4d1d7be119c1e306"} Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.685256 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.790891 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-config-data\") pod \"8110268a-391c-479e-bb86-3bf7fdd0c987\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.791051 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvwr4\" (UniqueName: \"kubernetes.io/projected/8110268a-391c-479e-bb86-3bf7fdd0c987-kube-api-access-qvwr4\") pod \"8110268a-391c-479e-bb86-3bf7fdd0c987\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.791123 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-scripts\") pod \"8110268a-391c-479e-bb86-3bf7fdd0c987\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.791239 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8110268a-391c-479e-bb86-3bf7fdd0c987-logs\") pod \"8110268a-391c-479e-bb86-3bf7fdd0c987\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.791326 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-combined-ca-bundle\") pod \"8110268a-391c-479e-bb86-3bf7fdd0c987\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.791397 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"8110268a-391c-479e-bb86-3bf7fdd0c987\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.791441 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8110268a-391c-479e-bb86-3bf7fdd0c987-httpd-run\") pod \"8110268a-391c-479e-bb86-3bf7fdd0c987\" (UID: \"8110268a-391c-479e-bb86-3bf7fdd0c987\") " Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.792241 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8110268a-391c-479e-bb86-3bf7fdd0c987-logs" (OuterVolumeSpecName: "logs") pod "8110268a-391c-479e-bb86-3bf7fdd0c987" (UID: "8110268a-391c-479e-bb86-3bf7fdd0c987"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.792684 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8110268a-391c-479e-bb86-3bf7fdd0c987-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8110268a-391c-479e-bb86-3bf7fdd0c987" (UID: "8110268a-391c-479e-bb86-3bf7fdd0c987"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.792980 4828 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8110268a-391c-479e-bb86-3bf7fdd0c987-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.793004 4828 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8110268a-391c-479e-bb86-3bf7fdd0c987-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.798045 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "8110268a-391c-479e-bb86-3bf7fdd0c987" (UID: "8110268a-391c-479e-bb86-3bf7fdd0c987"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.796466 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8110268a-391c-479e-bb86-3bf7fdd0c987-kube-api-access-qvwr4" (OuterVolumeSpecName: "kube-api-access-qvwr4") pod "8110268a-391c-479e-bb86-3bf7fdd0c987" (UID: "8110268a-391c-479e-bb86-3bf7fdd0c987"). InnerVolumeSpecName "kube-api-access-qvwr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.816529 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-scripts" (OuterVolumeSpecName: "scripts") pod "8110268a-391c-479e-bb86-3bf7fdd0c987" (UID: "8110268a-391c-479e-bb86-3bf7fdd0c987"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.826330 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8110268a-391c-479e-bb86-3bf7fdd0c987" (UID: "8110268a-391c-479e-bb86-3bf7fdd0c987"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.858039 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-config-data" (OuterVolumeSpecName: "config-data") pod "8110268a-391c-479e-bb86-3bf7fdd0c987" (UID: "8110268a-391c-479e-bb86-3bf7fdd0c987"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.895605 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.895642 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvwr4\" (UniqueName: \"kubernetes.io/projected/8110268a-391c-479e-bb86-3bf7fdd0c987-kube-api-access-qvwr4\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.895654 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.895664 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8110268a-391c-479e-bb86-3bf7fdd0c987-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.895686 4828 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.921940 4828 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 10 19:20:07 crc kubenswrapper[4828]: I1210 19:20:07.999068 4828 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.066851 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.235597 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/832260f7-4d4e-48ed-a72c-af9d73156d70-httpd-run\") pod \"832260f7-4d4e-48ed-a72c-af9d73156d70\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.235767 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"832260f7-4d4e-48ed-a72c-af9d73156d70\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.236026 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-scripts\") pod \"832260f7-4d4e-48ed-a72c-af9d73156d70\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.236132 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbpg2\" (UniqueName: \"kubernetes.io/projected/832260f7-4d4e-48ed-a72c-af9d73156d70-kube-api-access-hbpg2\") pod \"832260f7-4d4e-48ed-a72c-af9d73156d70\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.236280 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/832260f7-4d4e-48ed-a72c-af9d73156d70-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "832260f7-4d4e-48ed-a72c-af9d73156d70" (UID: "832260f7-4d4e-48ed-a72c-af9d73156d70"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.236405 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-combined-ca-bundle\") pod \"832260f7-4d4e-48ed-a72c-af9d73156d70\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.236530 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-config-data\") pod \"832260f7-4d4e-48ed-a72c-af9d73156d70\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.236656 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/832260f7-4d4e-48ed-a72c-af9d73156d70-logs\") pod \"832260f7-4d4e-48ed-a72c-af9d73156d70\" (UID: \"832260f7-4d4e-48ed-a72c-af9d73156d70\") " Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.237538 4828 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/832260f7-4d4e-48ed-a72c-af9d73156d70-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.242504 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/832260f7-4d4e-48ed-a72c-af9d73156d70-logs" (OuterVolumeSpecName: "logs") pod "832260f7-4d4e-48ed-a72c-af9d73156d70" (UID: "832260f7-4d4e-48ed-a72c-af9d73156d70"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.251770 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "832260f7-4d4e-48ed-a72c-af9d73156d70" (UID: "832260f7-4d4e-48ed-a72c-af9d73156d70"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.252109 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-scripts" (OuterVolumeSpecName: "scripts") pod "832260f7-4d4e-48ed-a72c-af9d73156d70" (UID: "832260f7-4d4e-48ed-a72c-af9d73156d70"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.255640 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/832260f7-4d4e-48ed-a72c-af9d73156d70-kube-api-access-hbpg2" (OuterVolumeSpecName: "kube-api-access-hbpg2") pod "832260f7-4d4e-48ed-a72c-af9d73156d70" (UID: "832260f7-4d4e-48ed-a72c-af9d73156d70"). InnerVolumeSpecName "kube-api-access-hbpg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.270648 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.271277 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"832260f7-4d4e-48ed-a72c-af9d73156d70","Type":"ContainerDied","Data":"82192848e612c9d7dd68570ea2e89696c016753b7cc32c1b59e46bf336b84ebd"} Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.271349 4828 scope.go:117] "RemoveContainer" containerID="2034e9c324dd4b2994000f8e4b97e94ee61931119f5b59356d6f9a63990e46a1" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.278748 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8110268a-391c-479e-bb86-3bf7fdd0c987","Type":"ContainerDied","Data":"e7606cb9ed8f8217944bd6b88f3a2c73d0e392c650bbd86c69bfb3b2a9fafb6c"} Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.278831 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.289393 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "832260f7-4d4e-48ed-a72c-af9d73156d70" (UID: "832260f7-4d4e-48ed-a72c-af9d73156d70"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.293203 4828 generic.go:334] "Generic (PLEG): container finished" podID="1a634b8e-6ee2-4830-a764-2fcc5300b15d" containerID="da2f7b7fd47c67b352b34c029147284bbad04762c536da585a6f875dbf0617d8" exitCode=0 Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.293244 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2s89p" event={"ID":"1a634b8e-6ee2-4830-a764-2fcc5300b15d","Type":"ContainerDied","Data":"da2f7b7fd47c67b352b34c029147284bbad04762c536da585a6f875dbf0617d8"} Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.312305 4828 scope.go:117] "RemoveContainer" containerID="0b860ef7ec03f62883697c0aeab5dbc5fef529e10b76525a4d1d7be119c1e306" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.352601 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.352672 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbpg2\" (UniqueName: \"kubernetes.io/projected/832260f7-4d4e-48ed-a72c-af9d73156d70-kube-api-access-hbpg2\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.352685 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.352693 4828 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/832260f7-4d4e-48ed-a72c-af9d73156d70-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.352723 4828 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.352928 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-config-data" (OuterVolumeSpecName: "config-data") pod "832260f7-4d4e-48ed-a72c-af9d73156d70" (UID: "832260f7-4d4e-48ed-a72c-af9d73156d70"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.370581 4828 scope.go:117] "RemoveContainer" containerID="35a12ed90ec6d821670de8a0a219594bf0d6168162509a1b23e43b4bcfbbbd66" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.390691 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.402528 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.410362 4828 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.412999 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:20:08 crc kubenswrapper[4828]: E1210 19:20:08.413658 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8110268a-391c-479e-bb86-3bf7fdd0c987" containerName="glance-httpd" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.413678 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="8110268a-391c-479e-bb86-3bf7fdd0c987" containerName="glance-httpd" Dec 10 19:20:08 crc kubenswrapper[4828]: E1210 19:20:08.413701 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="832260f7-4d4e-48ed-a72c-af9d73156d70" containerName="glance-httpd" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.413710 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="832260f7-4d4e-48ed-a72c-af9d73156d70" containerName="glance-httpd" Dec 10 19:20:08 crc kubenswrapper[4828]: E1210 19:20:08.413750 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8110268a-391c-479e-bb86-3bf7fdd0c987" containerName="glance-log" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.413758 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="8110268a-391c-479e-bb86-3bf7fdd0c987" containerName="glance-log" Dec 10 19:20:08 crc kubenswrapper[4828]: E1210 19:20:08.413779 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="832260f7-4d4e-48ed-a72c-af9d73156d70" containerName="glance-log" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.413788 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="832260f7-4d4e-48ed-a72c-af9d73156d70" containerName="glance-log" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.414086 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="8110268a-391c-479e-bb86-3bf7fdd0c987" containerName="glance-log" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.414105 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="832260f7-4d4e-48ed-a72c-af9d73156d70" containerName="glance-log" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.414115 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="832260f7-4d4e-48ed-a72c-af9d73156d70" containerName="glance-httpd" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.414128 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="8110268a-391c-479e-bb86-3bf7fdd0c987" containerName="glance-httpd" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.418191 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.420535 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.426830 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.429412 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.431536 4828 scope.go:117] "RemoveContainer" containerID="f7cf4cd4cf43f81e166d49840007a9912685de8a42e6f2611af2b924d4af86e3" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.456833 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.456988 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.457101 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9llcc\" (UniqueName: \"kubernetes.io/projected/c5296458-e976-4d87-a73f-3fb10cddd3d0-kube-api-access-9llcc\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.457201 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-scripts\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.457385 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-config-data\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.457531 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5296458-e976-4d87-a73f-3fb10cddd3d0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.457697 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5296458-e976-4d87-a73f-3fb10cddd3d0-logs\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.457952 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.458210 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/832260f7-4d4e-48ed-a72c-af9d73156d70-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.458296 4828 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.560081 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-config-data\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.560478 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5296458-e976-4d87-a73f-3fb10cddd3d0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.560521 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5296458-e976-4d87-a73f-3fb10cddd3d0-logs\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.560565 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.560715 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.560740 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.560765 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9llcc\" (UniqueName: \"kubernetes.io/projected/c5296458-e976-4d87-a73f-3fb10cddd3d0-kube-api-access-9llcc\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.560786 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-scripts\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.560999 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5296458-e976-4d87-a73f-3fb10cddd3d0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.561220 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.562678 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5296458-e976-4d87-a73f-3fb10cddd3d0-logs\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.564704 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-scripts\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.565333 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.570341 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-config-data\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.575858 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.579493 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9llcc\" (UniqueName: \"kubernetes.io/projected/c5296458-e976-4d87-a73f-3fb10cddd3d0-kube-api-access-9llcc\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.593046 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.653832 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-ptl7r"] Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.739411 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.872075 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-lh7pb" podUID="8e282513-5ba6-4eec-8738-8406c723a69b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.146:5353: i/o timeout" Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.944473 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.968852 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.992751 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:20:08 crc kubenswrapper[4828]: I1210 19:20:08.994640 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.009123 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.009327 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.020274 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.072665 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.072980 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.073042 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.073090 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75ppq\" (UniqueName: \"kubernetes.io/projected/53369a6e-a6da-4d63-aca6-e1bec8ab047e-kube-api-access-75ppq\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.073114 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53369a6e-a6da-4d63-aca6-e1bec8ab047e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.073166 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.073184 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.073350 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53369a6e-a6da-4d63-aca6-e1bec8ab047e-logs\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.175521 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75ppq\" (UniqueName: \"kubernetes.io/projected/53369a6e-a6da-4d63-aca6-e1bec8ab047e-kube-api-access-75ppq\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.175575 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53369a6e-a6da-4d63-aca6-e1bec8ab047e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.175625 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.175641 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.175679 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53369a6e-a6da-4d63-aca6-e1bec8ab047e-logs\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.175771 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.175819 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.175860 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.177358 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53369a6e-a6da-4d63-aca6-e1bec8ab047e-logs\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.177555 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.177672 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53369a6e-a6da-4d63-aca6-e1bec8ab047e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.182519 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.185558 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.188659 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.188842 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.200844 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75ppq\" (UniqueName: \"kubernetes.io/projected/53369a6e-a6da-4d63-aca6-e1bec8ab047e-kube-api-access-75ppq\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.219497 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.338565 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.382055 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cdea527-5341-46d2-b11f-9c88c4b35469","Type":"ContainerStarted","Data":"c00655fab2f22d8330b86cf36e17d7861f6fd58142cbd516fb930de7f29f0494"} Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.384315 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a" containerID="fc6c64ed594fe07d98e5a50b61dc2fe5a667292c054740f5de9c8d1fe0244a74" exitCode=0 Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.384371 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gwbxv" event={"ID":"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a","Type":"ContainerDied","Data":"fc6c64ed594fe07d98e5a50b61dc2fe5a667292c054740f5de9c8d1fe0244a74"} Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.396821 4828 generic.go:334] "Generic (PLEG): container finished" podID="046c39f9-a9eb-45d4-9965-5575059fdcfc" containerID="331a9ca0786faa74ad0eec3f00586a0ba2d92d7219a89fcc0d0fa25ae53c366e" exitCode=0 Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.397350 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" event={"ID":"046c39f9-a9eb-45d4-9965-5575059fdcfc","Type":"ContainerDied","Data":"331a9ca0786faa74ad0eec3f00586a0ba2d92d7219a89fcc0d0fa25ae53c366e"} Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.397428 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" event={"ID":"046c39f9-a9eb-45d4-9965-5575059fdcfc","Type":"ContainerStarted","Data":"c5f38f0df1bf1b2b671bcab7088e311c6d38a554cc4655fc88651abac0db7265"} Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.398376 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.837211 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8110268a-391c-479e-bb86-3bf7fdd0c987" path="/var/lib/kubelet/pods/8110268a-391c-479e-bb86-3bf7fdd0c987/volumes" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.838351 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="832260f7-4d4e-48ed-a72c-af9d73156d70" path="/var/lib/kubelet/pods/832260f7-4d4e-48ed-a72c-af9d73156d70/volumes" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.869674 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2s89p" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.895156 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1a634b8e-6ee2-4830-a764-2fcc5300b15d-db-sync-config-data\") pod \"1a634b8e-6ee2-4830-a764-2fcc5300b15d\" (UID: \"1a634b8e-6ee2-4830-a764-2fcc5300b15d\") " Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.895391 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzt9p\" (UniqueName: \"kubernetes.io/projected/1a634b8e-6ee2-4830-a764-2fcc5300b15d-kube-api-access-qzt9p\") pod \"1a634b8e-6ee2-4830-a764-2fcc5300b15d\" (UID: \"1a634b8e-6ee2-4830-a764-2fcc5300b15d\") " Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.895497 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a634b8e-6ee2-4830-a764-2fcc5300b15d-combined-ca-bundle\") pod \"1a634b8e-6ee2-4830-a764-2fcc5300b15d\" (UID: \"1a634b8e-6ee2-4830-a764-2fcc5300b15d\") " Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.905765 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a634b8e-6ee2-4830-a764-2fcc5300b15d-kube-api-access-qzt9p" (OuterVolumeSpecName: "kube-api-access-qzt9p") pod "1a634b8e-6ee2-4830-a764-2fcc5300b15d" (UID: "1a634b8e-6ee2-4830-a764-2fcc5300b15d"). InnerVolumeSpecName "kube-api-access-qzt9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.908446 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a634b8e-6ee2-4830-a764-2fcc5300b15d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1a634b8e-6ee2-4830-a764-2fcc5300b15d" (UID: "1a634b8e-6ee2-4830-a764-2fcc5300b15d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:09 crc kubenswrapper[4828]: I1210 19:20:09.938460 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a634b8e-6ee2-4830-a764-2fcc5300b15d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a634b8e-6ee2-4830-a764-2fcc5300b15d" (UID: "1a634b8e-6ee2-4830-a764-2fcc5300b15d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.000370 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzt9p\" (UniqueName: \"kubernetes.io/projected/1a634b8e-6ee2-4830-a764-2fcc5300b15d-kube-api-access-qzt9p\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.000405 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a634b8e-6ee2-4830-a764-2fcc5300b15d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.000416 4828 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1a634b8e-6ee2-4830-a764-2fcc5300b15d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.074109 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:20:10 crc kubenswrapper[4828]: W1210 19:20:10.079174 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53369a6e_a6da_4d63_aca6_e1bec8ab047e.slice/crio-22aa1decab2296eb980332fae51e49edc430a14cb1184cbbde9de15eed0dc3e6 WatchSource:0}: Error finding container 22aa1decab2296eb980332fae51e49edc430a14cb1184cbbde9de15eed0dc3e6: Status 404 returned error can't find the container with id 22aa1decab2296eb980332fae51e49edc430a14cb1184cbbde9de15eed0dc3e6 Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.472303 4828 generic.go:334] "Generic (PLEG): container finished" podID="a56b6762-0c16-4b35-a11b-16e34810aeab" containerID="c5f3ce88ff6d90ede4552698a8c5ec775f1d143ab5228560c41b371330fc2ff1" exitCode=0 Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.472393 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-n7jdh" event={"ID":"a56b6762-0c16-4b35-a11b-16e34810aeab","Type":"ContainerDied","Data":"c5f3ce88ff6d90ede4552698a8c5ec775f1d143ab5228560c41b371330fc2ff1"} Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.485532 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5296458-e976-4d87-a73f-3fb10cddd3d0","Type":"ContainerStarted","Data":"4655737d2ef0ac6fce17b004ea4fb37aeb4e7c011d3fb75feae97da21930b36a"} Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.487540 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5296458-e976-4d87-a73f-3fb10cddd3d0","Type":"ContainerStarted","Data":"4a451e3223768c991f379f410c411c08c1783117a52e4d37b5e4cbd9f7adc27e"} Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.493878 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53369a6e-a6da-4d63-aca6-e1bec8ab047e","Type":"ContainerStarted","Data":"22aa1decab2296eb980332fae51e49edc430a14cb1184cbbde9de15eed0dc3e6"} Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.505818 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-2s89p" event={"ID":"1a634b8e-6ee2-4830-a764-2fcc5300b15d","Type":"ContainerDied","Data":"3add87f7232f1c5abc63d401be94d5333934f93b421e756db28435328e6f711d"} Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.505874 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3add87f7232f1c5abc63d401be94d5333934f93b421e756db28435328e6f711d" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.505931 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-2s89p" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.556664 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" event={"ID":"046c39f9-a9eb-45d4-9965-5575059fdcfc","Type":"ContainerStarted","Data":"4df44e548c885af1b534cabea4024ab283f30c303cf6d212e6f18165efe76efe"} Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.556835 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.628970 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-69884f6747-2l55b"] Dec 10 19:20:10 crc kubenswrapper[4828]: E1210 19:20:10.629578 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a634b8e-6ee2-4830-a764-2fcc5300b15d" containerName="barbican-db-sync" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.629605 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a634b8e-6ee2-4830-a764-2fcc5300b15d" containerName="barbican-db-sync" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.629871 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a634b8e-6ee2-4830-a764-2fcc5300b15d" containerName="barbican-db-sync" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.631336 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-69884f6747-2l55b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.643331 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.644026 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-gcmgv" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.644269 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.695347 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-85965bf486-fgbxr"] Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.697614 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.710199 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.729864 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv9h5\" (UniqueName: \"kubernetes.io/projected/1bc1ebd1-1426-4def-a07d-5b0b561f2082-kube-api-access-lv9h5\") pod \"barbican-keystone-listener-85965bf486-fgbxr\" (UID: \"1bc1ebd1-1426-4def-a07d-5b0b561f2082\") " pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.729934 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bc1ebd1-1426-4def-a07d-5b0b561f2082-logs\") pod \"barbican-keystone-listener-85965bf486-fgbxr\" (UID: \"1bc1ebd1-1426-4def-a07d-5b0b561f2082\") " pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.729963 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bc1ebd1-1426-4def-a07d-5b0b561f2082-config-data-custom\") pod \"barbican-keystone-listener-85965bf486-fgbxr\" (UID: \"1bc1ebd1-1426-4def-a07d-5b0b561f2082\") " pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.729982 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc1ebd1-1426-4def-a07d-5b0b561f2082-combined-ca-bundle\") pod \"barbican-keystone-listener-85965bf486-fgbxr\" (UID: \"1bc1ebd1-1426-4def-a07d-5b0b561f2082\") " pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.730052 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee9fcef-2371-4614-b67f-6cef46b20c6f-combined-ca-bundle\") pod \"barbican-worker-69884f6747-2l55b\" (UID: \"7ee9fcef-2371-4614-b67f-6cef46b20c6f\") " pod="openstack/barbican-worker-69884f6747-2l55b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.730237 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcgsb\" (UniqueName: \"kubernetes.io/projected/7ee9fcef-2371-4614-b67f-6cef46b20c6f-kube-api-access-xcgsb\") pod \"barbican-worker-69884f6747-2l55b\" (UID: \"7ee9fcef-2371-4614-b67f-6cef46b20c6f\") " pod="openstack/barbican-worker-69884f6747-2l55b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.730278 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee9fcef-2371-4614-b67f-6cef46b20c6f-config-data\") pod \"barbican-worker-69884f6747-2l55b\" (UID: \"7ee9fcef-2371-4614-b67f-6cef46b20c6f\") " pod="openstack/barbican-worker-69884f6747-2l55b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.730461 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc1ebd1-1426-4def-a07d-5b0b561f2082-config-data\") pod \"barbican-keystone-listener-85965bf486-fgbxr\" (UID: \"1bc1ebd1-1426-4def-a07d-5b0b561f2082\") " pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.730522 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7ee9fcef-2371-4614-b67f-6cef46b20c6f-config-data-custom\") pod \"barbican-worker-69884f6747-2l55b\" (UID: \"7ee9fcef-2371-4614-b67f-6cef46b20c6f\") " pod="openstack/barbican-worker-69884f6747-2l55b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.730565 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ee9fcef-2371-4614-b67f-6cef46b20c6f-logs\") pod \"barbican-worker-69884f6747-2l55b\" (UID: \"7ee9fcef-2371-4614-b67f-6cef46b20c6f\") " pod="openstack/barbican-worker-69884f6747-2l55b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.739654 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-85965bf486-fgbxr"] Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.784060 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" podStartSLOduration=5.784041968 podStartE2EDuration="5.784041968s" podCreationTimestamp="2025-12-10 19:20:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:10.619235708 +0000 UTC m=+1491.129846723" watchObservedRunningTime="2025-12-10 19:20:10.784041968 +0000 UTC m=+1491.294652973" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.786152 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-69884f6747-2l55b"] Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.832148 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv9h5\" (UniqueName: \"kubernetes.io/projected/1bc1ebd1-1426-4def-a07d-5b0b561f2082-kube-api-access-lv9h5\") pod \"barbican-keystone-listener-85965bf486-fgbxr\" (UID: \"1bc1ebd1-1426-4def-a07d-5b0b561f2082\") " pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.832211 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bc1ebd1-1426-4def-a07d-5b0b561f2082-logs\") pod \"barbican-keystone-listener-85965bf486-fgbxr\" (UID: \"1bc1ebd1-1426-4def-a07d-5b0b561f2082\") " pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.832241 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bc1ebd1-1426-4def-a07d-5b0b561f2082-config-data-custom\") pod \"barbican-keystone-listener-85965bf486-fgbxr\" (UID: \"1bc1ebd1-1426-4def-a07d-5b0b561f2082\") " pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.832265 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc1ebd1-1426-4def-a07d-5b0b561f2082-combined-ca-bundle\") pod \"barbican-keystone-listener-85965bf486-fgbxr\" (UID: \"1bc1ebd1-1426-4def-a07d-5b0b561f2082\") " pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.832322 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee9fcef-2371-4614-b67f-6cef46b20c6f-combined-ca-bundle\") pod \"barbican-worker-69884f6747-2l55b\" (UID: \"7ee9fcef-2371-4614-b67f-6cef46b20c6f\") " pod="openstack/barbican-worker-69884f6747-2l55b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.832385 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcgsb\" (UniqueName: \"kubernetes.io/projected/7ee9fcef-2371-4614-b67f-6cef46b20c6f-kube-api-access-xcgsb\") pod \"barbican-worker-69884f6747-2l55b\" (UID: \"7ee9fcef-2371-4614-b67f-6cef46b20c6f\") " pod="openstack/barbican-worker-69884f6747-2l55b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.838306 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee9fcef-2371-4614-b67f-6cef46b20c6f-config-data\") pod \"barbican-worker-69884f6747-2l55b\" (UID: \"7ee9fcef-2371-4614-b67f-6cef46b20c6f\") " pod="openstack/barbican-worker-69884f6747-2l55b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.838494 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc1ebd1-1426-4def-a07d-5b0b561f2082-config-data\") pod \"barbican-keystone-listener-85965bf486-fgbxr\" (UID: \"1bc1ebd1-1426-4def-a07d-5b0b561f2082\") " pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.838553 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7ee9fcef-2371-4614-b67f-6cef46b20c6f-config-data-custom\") pod \"barbican-worker-69884f6747-2l55b\" (UID: \"7ee9fcef-2371-4614-b67f-6cef46b20c6f\") " pod="openstack/barbican-worker-69884f6747-2l55b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.838594 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ee9fcef-2371-4614-b67f-6cef46b20c6f-logs\") pod \"barbican-worker-69884f6747-2l55b\" (UID: \"7ee9fcef-2371-4614-b67f-6cef46b20c6f\") " pod="openstack/barbican-worker-69884f6747-2l55b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.839199 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7ee9fcef-2371-4614-b67f-6cef46b20c6f-logs\") pod \"barbican-worker-69884f6747-2l55b\" (UID: \"7ee9fcef-2371-4614-b67f-6cef46b20c6f\") " pod="openstack/barbican-worker-69884f6747-2l55b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.839880 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1bc1ebd1-1426-4def-a07d-5b0b561f2082-logs\") pod \"barbican-keystone-listener-85965bf486-fgbxr\" (UID: \"1bc1ebd1-1426-4def-a07d-5b0b561f2082\") " pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.859231 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bc1ebd1-1426-4def-a07d-5b0b561f2082-config-data-custom\") pod \"barbican-keystone-listener-85965bf486-fgbxr\" (UID: \"1bc1ebd1-1426-4def-a07d-5b0b561f2082\") " pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.859948 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc1ebd1-1426-4def-a07d-5b0b561f2082-combined-ca-bundle\") pod \"barbican-keystone-listener-85965bf486-fgbxr\" (UID: \"1bc1ebd1-1426-4def-a07d-5b0b561f2082\") " pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.863695 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-ptl7r"] Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.868909 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee9fcef-2371-4614-b67f-6cef46b20c6f-combined-ca-bundle\") pod \"barbican-worker-69884f6747-2l55b\" (UID: \"7ee9fcef-2371-4614-b67f-6cef46b20c6f\") " pod="openstack/barbican-worker-69884f6747-2l55b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.869101 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7ee9fcef-2371-4614-b67f-6cef46b20c6f-config-data-custom\") pod \"barbican-worker-69884f6747-2l55b\" (UID: \"7ee9fcef-2371-4614-b67f-6cef46b20c6f\") " pod="openstack/barbican-worker-69884f6747-2l55b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.871959 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv9h5\" (UniqueName: \"kubernetes.io/projected/1bc1ebd1-1426-4def-a07d-5b0b561f2082-kube-api-access-lv9h5\") pod \"barbican-keystone-listener-85965bf486-fgbxr\" (UID: \"1bc1ebd1-1426-4def-a07d-5b0b561f2082\") " pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.873534 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc1ebd1-1426-4def-a07d-5b0b561f2082-config-data\") pod \"barbican-keystone-listener-85965bf486-fgbxr\" (UID: \"1bc1ebd1-1426-4def-a07d-5b0b561f2082\") " pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.874096 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcgsb\" (UniqueName: \"kubernetes.io/projected/7ee9fcef-2371-4614-b67f-6cef46b20c6f-kube-api-access-xcgsb\") pod \"barbican-worker-69884f6747-2l55b\" (UID: \"7ee9fcef-2371-4614-b67f-6cef46b20c6f\") " pod="openstack/barbican-worker-69884f6747-2l55b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.879899 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ee9fcef-2371-4614-b67f-6cef46b20c6f-config-data\") pod \"barbican-worker-69884f6747-2l55b\" (UID: \"7ee9fcef-2371-4614-b67f-6cef46b20c6f\") " pod="openstack/barbican-worker-69884f6747-2l55b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.879974 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-c5xrq"] Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.882640 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.905107 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-c5xrq"] Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.918127 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7cf7687b58-b784b"] Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.920336 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.925776 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.928877 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7cf7687b58-b784b"] Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.945206 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-config-data\") pod \"barbican-api-7cf7687b58-b784b\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.945269 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-combined-ca-bundle\") pod \"barbican-api-7cf7687b58-b784b\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.945322 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-config\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.945347 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-ovsdbserver-sb\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.945382 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcxn9\" (UniqueName: \"kubernetes.io/projected/339f67ce-2474-4399-a2ad-2e4bcfdd01af-kube-api-access-fcxn9\") pod \"barbican-api-7cf7687b58-b784b\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.945408 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-ovsdbserver-nb\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.945425 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-config-data-custom\") pod \"barbican-api-7cf7687b58-b784b\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.945516 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/339f67ce-2474-4399-a2ad-2e4bcfdd01af-logs\") pod \"barbican-api-7cf7687b58-b784b\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.945563 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-dns-swift-storage-0\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.945621 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r29s\" (UniqueName: \"kubernetes.io/projected/00ad66a4-0ac5-4678-9d54-26738e3eb1af-kube-api-access-9r29s\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:10 crc kubenswrapper[4828]: I1210 19:20:10.945656 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-dns-svc\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.046082 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-69884f6747-2l55b" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.046439 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r29s\" (UniqueName: \"kubernetes.io/projected/00ad66a4-0ac5-4678-9d54-26738e3eb1af-kube-api-access-9r29s\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.046493 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-dns-svc\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.046551 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-config-data\") pod \"barbican-api-7cf7687b58-b784b\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.046591 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-combined-ca-bundle\") pod \"barbican-api-7cf7687b58-b784b\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.046633 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-config\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.046657 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-ovsdbserver-sb\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.046687 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcxn9\" (UniqueName: \"kubernetes.io/projected/339f67ce-2474-4399-a2ad-2e4bcfdd01af-kube-api-access-fcxn9\") pod \"barbican-api-7cf7687b58-b784b\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.046711 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-ovsdbserver-nb\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.046732 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-config-data-custom\") pod \"barbican-api-7cf7687b58-b784b\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.046842 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/339f67ce-2474-4399-a2ad-2e4bcfdd01af-logs\") pod \"barbican-api-7cf7687b58-b784b\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.046901 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-dns-swift-storage-0\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.047576 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-dns-svc\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.047623 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-dns-swift-storage-0\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.055651 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-ovsdbserver-nb\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.062110 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-config\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.063240 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-ovsdbserver-sb\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.063407 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/339f67ce-2474-4399-a2ad-2e4bcfdd01af-logs\") pod \"barbican-api-7cf7687b58-b784b\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.073933 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r29s\" (UniqueName: \"kubernetes.io/projected/00ad66a4-0ac5-4678-9d54-26738e3eb1af-kube-api-access-9r29s\") pod \"dnsmasq-dns-7c67bffd47-c5xrq\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.074728 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-combined-ca-bundle\") pod \"barbican-api-7cf7687b58-b784b\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.074944 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-config-data\") pod \"barbican-api-7cf7687b58-b784b\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.075433 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.076282 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-config-data-custom\") pod \"barbican-api-7cf7687b58-b784b\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.104165 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcxn9\" (UniqueName: \"kubernetes.io/projected/339f67ce-2474-4399-a2ad-2e4bcfdd01af-kube-api-access-fcxn9\") pod \"barbican-api-7cf7687b58-b784b\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.210157 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.253705 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-combined-ca-bundle\") pod \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.254126 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-credential-keys\") pod \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.254195 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7h7k\" (UniqueName: \"kubernetes.io/projected/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-kube-api-access-t7h7k\") pod \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.254296 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-scripts\") pod \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.254378 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-config-data\") pod \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.254448 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-fernet-keys\") pod \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\" (UID: \"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a\") " Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.318829 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.339714 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.524098 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-kube-api-access-t7h7k" (OuterVolumeSpecName: "kube-api-access-t7h7k") pod "1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a" (UID: "1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a"). InnerVolumeSpecName "kube-api-access-t7h7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.524309 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a" (UID: "1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.524396 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-scripts" (OuterVolumeSpecName: "scripts") pod "1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a" (UID: "1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.529896 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a" (UID: "1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.570413 4828 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.570441 4828 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.570455 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7h7k\" (UniqueName: \"kubernetes.io/projected/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-kube-api-access-t7h7k\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.570468 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.625665 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a" (UID: "1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.630928 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-config-data" (OuterVolumeSpecName: "config-data") pod "1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a" (UID: "1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.637258 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6f465c7bf6-5f8bw"] Dec 10 19:20:11 crc kubenswrapper[4828]: E1210 19:20:11.642997 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a" containerName="keystone-bootstrap" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.643028 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a" containerName="keystone-bootstrap" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.643343 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a" containerName="keystone-bootstrap" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.647723 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.652127 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.652342 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.665285 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6f465c7bf6-5f8bw"] Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.683948 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-internal-tls-certs\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.684037 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-credential-keys\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.684155 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-fernet-keys\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.684288 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-combined-ca-bundle\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.684355 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-529df\" (UniqueName: \"kubernetes.io/projected/aaa7ff68-b5a6-4f74-9879-6a51c6016578-kube-api-access-529df\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.684459 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-public-tls-certs\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.684493 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-scripts\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.684605 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-config-data\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.684672 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.684687 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.701599 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53369a6e-a6da-4d63-aca6-e1bec8ab047e","Type":"ContainerStarted","Data":"af613c0a15c1188568299015ab706de182e2534b39204ffc56468699e1878789"} Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.733941 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-gwbxv" event={"ID":"1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a","Type":"ContainerDied","Data":"5722fad71fd4a0a6de3283be7eb278f4e60e48a95598f6cb0867cb45060e76d6"} Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.733989 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5722fad71fd4a0a6de3283be7eb278f4e60e48a95598f6cb0867cb45060e76d6" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.734220 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.787373 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-combined-ca-bundle\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.787429 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-529df\" (UniqueName: \"kubernetes.io/projected/aaa7ff68-b5a6-4f74-9879-6a51c6016578-kube-api-access-529df\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.787478 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-public-tls-certs\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.787497 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-scripts\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.787555 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-config-data\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.787578 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-internal-tls-certs\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.787602 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-credential-keys\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.787651 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-fernet-keys\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.797374 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-config-data\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.808498 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-combined-ca-bundle\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.815415 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-internal-tls-certs\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.823601 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-credential-keys\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.847054 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-public-tls-certs\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.847586 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-fernet-keys\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.848450 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-529df\" (UniqueName: \"kubernetes.io/projected/aaa7ff68-b5a6-4f74-9879-6a51c6016578-kube-api-access-529df\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.850215 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aaa7ff68-b5a6-4f74-9879-6a51c6016578-scripts\") pod \"keystone-6f465c7bf6-5f8bw\" (UID: \"aaa7ff68-b5a6-4f74-9879-6a51c6016578\") " pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:11 crc kubenswrapper[4828]: I1210 19:20:11.972377 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:12 crc kubenswrapper[4828]: E1210 19:20:12.204426 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f2eb8fc_198b_4cf5_bcc9_e47f35b6324a.slice/crio-5722fad71fd4a0a6de3283be7eb278f4e60e48a95598f6cb0867cb45060e76d6\": RecentStats: unable to find data in memory cache]" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.442413 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-69884f6747-2l55b"] Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.467940 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-n7jdh" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.503169 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-85965bf486-fgbxr"] Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.559636 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-c5xrq"] Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.613858 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l6qg\" (UniqueName: \"kubernetes.io/projected/a56b6762-0c16-4b35-a11b-16e34810aeab-kube-api-access-5l6qg\") pod \"a56b6762-0c16-4b35-a11b-16e34810aeab\" (UID: \"a56b6762-0c16-4b35-a11b-16e34810aeab\") " Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.613945 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a56b6762-0c16-4b35-a11b-16e34810aeab-config\") pod \"a56b6762-0c16-4b35-a11b-16e34810aeab\" (UID: \"a56b6762-0c16-4b35-a11b-16e34810aeab\") " Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.614061 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a56b6762-0c16-4b35-a11b-16e34810aeab-combined-ca-bundle\") pod \"a56b6762-0c16-4b35-a11b-16e34810aeab\" (UID: \"a56b6762-0c16-4b35-a11b-16e34810aeab\") " Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.667812 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a56b6762-0c16-4b35-a11b-16e34810aeab-kube-api-access-5l6qg" (OuterVolumeSpecName: "kube-api-access-5l6qg") pod "a56b6762-0c16-4b35-a11b-16e34810aeab" (UID: "a56b6762-0c16-4b35-a11b-16e34810aeab"). InnerVolumeSpecName "kube-api-access-5l6qg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.703020 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a56b6762-0c16-4b35-a11b-16e34810aeab-config" (OuterVolumeSpecName: "config") pod "a56b6762-0c16-4b35-a11b-16e34810aeab" (UID: "a56b6762-0c16-4b35-a11b-16e34810aeab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.705888 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-c5xrq"] Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.710145 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a56b6762-0c16-4b35-a11b-16e34810aeab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a56b6762-0c16-4b35-a11b-16e34810aeab" (UID: "a56b6762-0c16-4b35-a11b-16e34810aeab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.719369 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l6qg\" (UniqueName: \"kubernetes.io/projected/a56b6762-0c16-4b35-a11b-16e34810aeab-kube-api-access-5l6qg\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.722129 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-pxfhx"] Dec 10 19:20:12 crc kubenswrapper[4828]: E1210 19:20:12.722586 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a56b6762-0c16-4b35-a11b-16e34810aeab" containerName="neutron-db-sync" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.722600 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="a56b6762-0c16-4b35-a11b-16e34810aeab" containerName="neutron-db-sync" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.723155 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a56b6762-0c16-4b35-a11b-16e34810aeab-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.723951 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a56b6762-0c16-4b35-a11b-16e34810aeab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.744298 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="a56b6762-0c16-4b35-a11b-16e34810aeab" containerName="neutron-db-sync" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.752285 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.802932 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-pxfhx"] Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.846142 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7cf7687b58-b784b"] Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.861585 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" event={"ID":"00ad66a4-0ac5-4678-9d54-26738e3eb1af","Type":"ContainerStarted","Data":"6e790ee7dfd1ffe39373260f9dc5b64bb1aec5b33e30608d5762ff851a48da8d"} Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.892947 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-66756f49f6-r2lds"] Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.935027 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4jvx\" (UniqueName: \"kubernetes.io/projected/ef7ea83b-99e3-4460-b3db-63b0fae43070-kube-api-access-g4jvx\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.935205 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.935242 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-config\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.935296 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.935595 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.935632 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.936840 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-69884f6747-2l55b" event={"ID":"7ee9fcef-2371-4614-b67f-6cef46b20c6f","Type":"ContainerStarted","Data":"3f566bea27be6aa427b17dcdd49851c36df7080b2b566c7c6ab4c7d54011a50b"} Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.937026 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.945312 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.959248 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5296458-e976-4d87-a73f-3fb10cddd3d0","Type":"ContainerStarted","Data":"9213c851c5e3a0adbad70dfa3fe4d86132ade02da751d4a97de33b2f0f9a6555"} Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.970143 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" event={"ID":"1bc1ebd1-1426-4def-a07d-5b0b561f2082","Type":"ContainerStarted","Data":"c2d041c090026493604eb03c57333b98a73ac40fa67d762816063810b35ef076"} Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.993058 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-n7jdh" Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.993638 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" podUID="046c39f9-a9eb-45d4-9965-5575059fdcfc" containerName="dnsmasq-dns" containerID="cri-o://4df44e548c885af1b534cabea4024ab283f30c303cf6d212e6f18165efe76efe" gracePeriod=10 Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.993833 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-n7jdh" event={"ID":"a56b6762-0c16-4b35-a11b-16e34810aeab","Type":"ContainerDied","Data":"13654f6e07ca21e622fe41f927a38612118ee5833f606cacd4282906ae4cb5be"} Dec 10 19:20:12 crc kubenswrapper[4828]: I1210 19:20:12.993857 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13654f6e07ca21e622fe41f927a38612118ee5833f606cacd4282906ae4cb5be" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.003261 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66756f49f6-r2lds"] Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.026599 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.026579195 podStartE2EDuration="5.026579195s" podCreationTimestamp="2025-12-10 19:20:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:13.002922322 +0000 UTC m=+1493.513533327" watchObservedRunningTime="2025-12-10 19:20:13.026579195 +0000 UTC m=+1493.537190200" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.038451 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.039002 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-ovndb-tls-certs\") pod \"neutron-66756f49f6-r2lds\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.039118 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.039205 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.039325 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-httpd-config\") pod \"neutron-66756f49f6-r2lds\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.039403 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4jvx\" (UniqueName: \"kubernetes.io/projected/ef7ea83b-99e3-4460-b3db-63b0fae43070-kube-api-access-g4jvx\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.039489 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvc92\" (UniqueName: \"kubernetes.io/projected/ddf32ed2-2731-497c-a021-6a00f495265f-kube-api-access-xvc92\") pod \"neutron-66756f49f6-r2lds\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.039567 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-config\") pod \"neutron-66756f49f6-r2lds\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.039739 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.039903 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-config\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.039983 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-combined-ca-bundle\") pod \"neutron-66756f49f6-r2lds\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.044542 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.044577 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.046950 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.047142 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.047661 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-config\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.061526 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6f465c7bf6-5f8bw"] Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.080627 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4jvx\" (UniqueName: \"kubernetes.io/projected/ef7ea83b-99e3-4460-b3db-63b0fae43070-kube-api-access-g4jvx\") pod \"dnsmasq-dns-848cf88cfc-pxfhx\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.153862 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-ovndb-tls-certs\") pod \"neutron-66756f49f6-r2lds\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.154155 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-httpd-config\") pod \"neutron-66756f49f6-r2lds\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.154252 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvc92\" (UniqueName: \"kubernetes.io/projected/ddf32ed2-2731-497c-a021-6a00f495265f-kube-api-access-xvc92\") pod \"neutron-66756f49f6-r2lds\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.154302 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-config\") pod \"neutron-66756f49f6-r2lds\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.154423 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-combined-ca-bundle\") pod \"neutron-66756f49f6-r2lds\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.159423 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-httpd-config\") pod \"neutron-66756f49f6-r2lds\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.168484 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-combined-ca-bundle\") pod \"neutron-66756f49f6-r2lds\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.174688 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-ovndb-tls-certs\") pod \"neutron-66756f49f6-r2lds\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.175517 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-config\") pod \"neutron-66756f49f6-r2lds\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.187484 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvc92\" (UniqueName: \"kubernetes.io/projected/ddf32ed2-2731-497c-a021-6a00f495265f-kube-api-access-xvc92\") pod \"neutron-66756f49f6-r2lds\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.286814 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:13 crc kubenswrapper[4828]: I1210 19:20:13.318598 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:14 crc kubenswrapper[4828]: I1210 19:20:14.021295 4828 generic.go:334] "Generic (PLEG): container finished" podID="046c39f9-a9eb-45d4-9965-5575059fdcfc" containerID="4df44e548c885af1b534cabea4024ab283f30c303cf6d212e6f18165efe76efe" exitCode=0 Dec 10 19:20:14 crc kubenswrapper[4828]: I1210 19:20:14.021616 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" event={"ID":"046c39f9-a9eb-45d4-9965-5575059fdcfc","Type":"ContainerDied","Data":"4df44e548c885af1b534cabea4024ab283f30c303cf6d212e6f18165efe76efe"} Dec 10 19:20:14 crc kubenswrapper[4828]: I1210 19:20:14.023468 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7cf7687b58-b784b" event={"ID":"339f67ce-2474-4399-a2ad-2e4bcfdd01af","Type":"ContainerStarted","Data":"991d0d54d2e48afe042e90163f9bc64719691578f5964443c3d76fd7f2d8382e"} Dec 10 19:20:14 crc kubenswrapper[4828]: I1210 19:20:14.025880 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6f465c7bf6-5f8bw" event={"ID":"aaa7ff68-b5a6-4f74-9879-6a51c6016578","Type":"ContainerStarted","Data":"edb8bd260d062f017eaa7277bfd41531497fde8fed511e7ce438d211cf124fc8"} Dec 10 19:20:14 crc kubenswrapper[4828]: I1210 19:20:14.138398 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66756f49f6-r2lds"] Dec 10 19:20:14 crc kubenswrapper[4828]: I1210 19:20:14.376618 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-pxfhx"] Dec 10 19:20:15 crc kubenswrapper[4828]: I1210 19:20:15.038997 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53369a6e-a6da-4d63-aca6-e1bec8ab047e","Type":"ContainerStarted","Data":"23700b3985da02baa67603281a49c24c027d44b3fb0df6c6db157900b864bded"} Dec 10 19:20:15 crc kubenswrapper[4828]: I1210 19:20:15.041175 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7cf7687b58-b784b" event={"ID":"339f67ce-2474-4399-a2ad-2e4bcfdd01af","Type":"ContainerStarted","Data":"fc3ea5b4555c800a8ca31a9c32426a49d13c9822a1a69181482217051a4cde86"} Dec 10 19:20:15 crc kubenswrapper[4828]: I1210 19:20:15.043082 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6f465c7bf6-5f8bw" event={"ID":"aaa7ff68-b5a6-4f74-9879-6a51c6016578","Type":"ContainerStarted","Data":"9c10a213288d9380f59ca0a76047d105a258c0d0ce8e3de3533c458a3212b11f"} Dec 10 19:20:15 crc kubenswrapper[4828]: I1210 19:20:15.043197 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:15 crc kubenswrapper[4828]: I1210 19:20:15.044782 4828 generic.go:334] "Generic (PLEG): container finished" podID="00ad66a4-0ac5-4678-9d54-26738e3eb1af" containerID="75cfb358c76392b2d4e6c561fe78f4056bd3eb7f6a2498b6b261edd0fa58d233" exitCode=0 Dec 10 19:20:15 crc kubenswrapper[4828]: I1210 19:20:15.044835 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" event={"ID":"00ad66a4-0ac5-4678-9d54-26738e3eb1af","Type":"ContainerDied","Data":"75cfb358c76392b2d4e6c561fe78f4056bd3eb7f6a2498b6b261edd0fa58d233"} Dec 10 19:20:15 crc kubenswrapper[4828]: I1210 19:20:15.070934 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.070917303 podStartE2EDuration="7.070917303s" podCreationTimestamp="2025-12-10 19:20:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:15.059856142 +0000 UTC m=+1495.570467157" watchObservedRunningTime="2025-12-10 19:20:15.070917303 +0000 UTC m=+1495.581528308" Dec 10 19:20:15 crc kubenswrapper[4828]: I1210 19:20:15.091650 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6f465c7bf6-5f8bw" podStartSLOduration=4.091626199 podStartE2EDuration="4.091626199s" podCreationTimestamp="2025-12-10 19:20:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:15.083017892 +0000 UTC m=+1495.593628897" watchObservedRunningTime="2025-12-10 19:20:15.091626199 +0000 UTC m=+1495.602237214" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.195479 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-56c96d9477-vps8l"] Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.197606 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.201148 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.201190 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.221708 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-56c96d9477-vps8l"] Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.249182 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-config\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.249304 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-combined-ca-bundle\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.249359 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-public-tls-certs\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.249401 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-internal-tls-certs\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.249433 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-ovndb-tls-certs\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.249486 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cx52\" (UniqueName: \"kubernetes.io/projected/2b65ecea-c14f-4702-be01-8952fabfb1be-kube-api-access-8cx52\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.249539 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-httpd-config\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.351439 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-combined-ca-bundle\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.351936 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-public-tls-certs\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.351993 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-internal-tls-certs\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.352025 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-ovndb-tls-certs\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.352086 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cx52\" (UniqueName: \"kubernetes.io/projected/2b65ecea-c14f-4702-be01-8952fabfb1be-kube-api-access-8cx52\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.352141 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-httpd-config\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.352225 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-config\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.361624 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-config\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.363843 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-internal-tls-certs\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.365610 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-httpd-config\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.372552 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-ovndb-tls-certs\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.373141 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-combined-ca-bundle\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.386996 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2b65ecea-c14f-4702-be01-8952fabfb1be-public-tls-certs\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.392643 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cx52\" (UniqueName: \"kubernetes.io/projected/2b65ecea-c14f-4702-be01-8952fabfb1be-kube-api-access-8cx52\") pod \"neutron-56c96d9477-vps8l\" (UID: \"2b65ecea-c14f-4702-be01-8952fabfb1be\") " pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.479142 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.485766 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.521753 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.558899 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-ovsdbserver-nb\") pod \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.558997 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-dns-swift-storage-0\") pod \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.559075 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-dns-svc\") pod \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.559132 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-ovsdbserver-nb\") pod \"046c39f9-a9eb-45d4-9965-5575059fdcfc\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.559179 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-ovsdbserver-sb\") pod \"046c39f9-a9eb-45d4-9965-5575059fdcfc\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.559220 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-dns-swift-storage-0\") pod \"046c39f9-a9eb-45d4-9965-5575059fdcfc\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.559236 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9r29s\" (UniqueName: \"kubernetes.io/projected/00ad66a4-0ac5-4678-9d54-26738e3eb1af-kube-api-access-9r29s\") pod \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.559289 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-config\") pod \"046c39f9-a9eb-45d4-9965-5575059fdcfc\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.559304 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lq76g\" (UniqueName: \"kubernetes.io/projected/046c39f9-a9eb-45d4-9965-5575059fdcfc-kube-api-access-lq76g\") pod \"046c39f9-a9eb-45d4-9965-5575059fdcfc\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.559323 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-dns-svc\") pod \"046c39f9-a9eb-45d4-9965-5575059fdcfc\" (UID: \"046c39f9-a9eb-45d4-9965-5575059fdcfc\") " Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.559375 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-ovsdbserver-sb\") pod \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.559438 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-config\") pod \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\" (UID: \"00ad66a4-0ac5-4678-9d54-26738e3eb1af\") " Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.659183 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00ad66a4-0ac5-4678-9d54-26738e3eb1af-kube-api-access-9r29s" (OuterVolumeSpecName: "kube-api-access-9r29s") pod "00ad66a4-0ac5-4678-9d54-26738e3eb1af" (UID: "00ad66a4-0ac5-4678-9d54-26738e3eb1af"). InnerVolumeSpecName "kube-api-access-9r29s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.661943 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9r29s\" (UniqueName: \"kubernetes.io/projected/00ad66a4-0ac5-4678-9d54-26738e3eb1af-kube-api-access-9r29s\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.686978 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/046c39f9-a9eb-45d4-9965-5575059fdcfc-kube-api-access-lq76g" (OuterVolumeSpecName: "kube-api-access-lq76g") pod "046c39f9-a9eb-45d4-9965-5575059fdcfc" (UID: "046c39f9-a9eb-45d4-9965-5575059fdcfc"). InnerVolumeSpecName "kube-api-access-lq76g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.734246 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-config" (OuterVolumeSpecName: "config") pod "00ad66a4-0ac5-4678-9d54-26738e3eb1af" (UID: "00ad66a4-0ac5-4678-9d54-26738e3eb1af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.761024 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "00ad66a4-0ac5-4678-9d54-26738e3eb1af" (UID: "00ad66a4-0ac5-4678-9d54-26738e3eb1af"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.768391 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lq76g\" (UniqueName: \"kubernetes.io/projected/046c39f9-a9eb-45d4-9965-5575059fdcfc-kube-api-access-lq76g\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.768427 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.768437 4828 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.769055 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "00ad66a4-0ac5-4678-9d54-26738e3eb1af" (UID: "00ad66a4-0ac5-4678-9d54-26738e3eb1af"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.775342 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "00ad66a4-0ac5-4678-9d54-26738e3eb1af" (UID: "00ad66a4-0ac5-4678-9d54-26738e3eb1af"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.787900 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-config" (OuterVolumeSpecName: "config") pod "046c39f9-a9eb-45d4-9965-5575059fdcfc" (UID: "046c39f9-a9eb-45d4-9965-5575059fdcfc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.792974 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "046c39f9-a9eb-45d4-9965-5575059fdcfc" (UID: "046c39f9-a9eb-45d4-9965-5575059fdcfc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.801349 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "00ad66a4-0ac5-4678-9d54-26738e3eb1af" (UID: "00ad66a4-0ac5-4678-9d54-26738e3eb1af"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.814844 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "046c39f9-a9eb-45d4-9965-5575059fdcfc" (UID: "046c39f9-a9eb-45d4-9965-5575059fdcfc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.824699 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "046c39f9-a9eb-45d4-9965-5575059fdcfc" (UID: "046c39f9-a9eb-45d4-9965-5575059fdcfc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.828235 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "046c39f9-a9eb-45d4-9965-5575059fdcfc" (UID: "046c39f9-a9eb-45d4-9965-5575059fdcfc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.870663 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.870706 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.870716 4828 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/00ad66a4-0ac5-4678-9d54-26738e3eb1af-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.870726 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.870736 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.870749 4828 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.870759 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:16 crc kubenswrapper[4828]: I1210 19:20:16.870769 4828 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/046c39f9-a9eb-45d4-9965-5575059fdcfc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.099291 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" event={"ID":"046c39f9-a9eb-45d4-9965-5575059fdcfc","Type":"ContainerDied","Data":"c5f38f0df1bf1b2b671bcab7088e311c6d38a554cc4655fc88651abac0db7265"} Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.099344 4828 scope.go:117] "RemoveContainer" containerID="4df44e548c885af1b534cabea4024ab283f30c303cf6d212e6f18165efe76efe" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.099359 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.101966 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" event={"ID":"00ad66a4-0ac5-4678-9d54-26738e3eb1af","Type":"ContainerDied","Data":"6e790ee7dfd1ffe39373260f9dc5b64bb1aec5b33e30608d5762ff851a48da8d"} Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.102023 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-c5xrq" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.176753 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-c5xrq"] Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.219645 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-c5xrq"] Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.237966 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-ptl7r"] Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.252736 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-ptl7r"] Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.669794 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6598f5c59d-btf5k"] Dec 10 19:20:17 crc kubenswrapper[4828]: E1210 19:20:17.670655 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00ad66a4-0ac5-4678-9d54-26738e3eb1af" containerName="init" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.670681 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="00ad66a4-0ac5-4678-9d54-26738e3eb1af" containerName="init" Dec 10 19:20:17 crc kubenswrapper[4828]: E1210 19:20:17.670733 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="046c39f9-a9eb-45d4-9965-5575059fdcfc" containerName="dnsmasq-dns" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.670742 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="046c39f9-a9eb-45d4-9965-5575059fdcfc" containerName="dnsmasq-dns" Dec 10 19:20:17 crc kubenswrapper[4828]: E1210 19:20:17.670764 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="046c39f9-a9eb-45d4-9965-5575059fdcfc" containerName="init" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.670773 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="046c39f9-a9eb-45d4-9965-5575059fdcfc" containerName="init" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.671058 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="00ad66a4-0ac5-4678-9d54-26738e3eb1af" containerName="init" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.671097 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="046c39f9-a9eb-45d4-9965-5575059fdcfc" containerName="dnsmasq-dns" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.672620 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.676682 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.676863 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.684994 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6598f5c59d-btf5k"] Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.795300 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d425401e-33bc-4dc1-adce-44622ce940df-config-data-custom\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.795400 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d425401e-33bc-4dc1-adce-44622ce940df-internal-tls-certs\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.795436 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d425401e-33bc-4dc1-adce-44622ce940df-logs\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.795493 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hslpp\" (UniqueName: \"kubernetes.io/projected/d425401e-33bc-4dc1-adce-44622ce940df-kube-api-access-hslpp\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.795541 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d425401e-33bc-4dc1-adce-44622ce940df-config-data\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.795601 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d425401e-33bc-4dc1-adce-44622ce940df-public-tls-certs\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.795662 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d425401e-33bc-4dc1-adce-44622ce940df-combined-ca-bundle\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.804942 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00ad66a4-0ac5-4678-9d54-26738e3eb1af" path="/var/lib/kubelet/pods/00ad66a4-0ac5-4678-9d54-26738e3eb1af/volumes" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.805458 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="046c39f9-a9eb-45d4-9965-5575059fdcfc" path="/var/lib/kubelet/pods/046c39f9-a9eb-45d4-9965-5575059fdcfc/volumes" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.897922 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d425401e-33bc-4dc1-adce-44622ce940df-config-data\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.898030 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d425401e-33bc-4dc1-adce-44622ce940df-public-tls-certs\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.898117 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d425401e-33bc-4dc1-adce-44622ce940df-combined-ca-bundle\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.898208 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d425401e-33bc-4dc1-adce-44622ce940df-config-data-custom\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.898245 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d425401e-33bc-4dc1-adce-44622ce940df-internal-tls-certs\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.898265 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d425401e-33bc-4dc1-adce-44622ce940df-logs\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.898311 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hslpp\" (UniqueName: \"kubernetes.io/projected/d425401e-33bc-4dc1-adce-44622ce940df-kube-api-access-hslpp\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:17 crc kubenswrapper[4828]: I1210 19:20:17.997741 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d425401e-33bc-4dc1-adce-44622ce940df-public-tls-certs\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:17.999076 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d425401e-33bc-4dc1-adce-44622ce940df-config-data\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.000133 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d425401e-33bc-4dc1-adce-44622ce940df-logs\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.002051 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d425401e-33bc-4dc1-adce-44622ce940df-internal-tls-certs\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.004380 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d425401e-33bc-4dc1-adce-44622ce940df-config-data-custom\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.004661 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d425401e-33bc-4dc1-adce-44622ce940df-combined-ca-bundle\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.016911 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hslpp\" (UniqueName: \"kubernetes.io/projected/d425401e-33bc-4dc1-adce-44622ce940df-kube-api-access-hslpp\") pod \"barbican-api-6598f5c59d-btf5k\" (UID: \"d425401e-33bc-4dc1-adce-44622ce940df\") " pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.044973 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xmnbf"] Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.047394 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.057970 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xmnbf"] Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.204918 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-catalog-content\") pod \"redhat-operators-xmnbf\" (UID: \"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a\") " pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.204999 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p97dz\" (UniqueName: \"kubernetes.io/projected/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-kube-api-access-p97dz\") pod \"redhat-operators-xmnbf\" (UID: \"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a\") " pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.205055 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-utilities\") pod \"redhat-operators-xmnbf\" (UID: \"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a\") " pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.294159 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.307041 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p97dz\" (UniqueName: \"kubernetes.io/projected/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-kube-api-access-p97dz\") pod \"redhat-operators-xmnbf\" (UID: \"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a\") " pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.307145 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-utilities\") pod \"redhat-operators-xmnbf\" (UID: \"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a\") " pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.307350 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-catalog-content\") pod \"redhat-operators-xmnbf\" (UID: \"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a\") " pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.307936 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-catalog-content\") pod \"redhat-operators-xmnbf\" (UID: \"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a\") " pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.308200 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-utilities\") pod \"redhat-operators-xmnbf\" (UID: \"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a\") " pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.341252 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p97dz\" (UniqueName: \"kubernetes.io/projected/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-kube-api-access-p97dz\") pod \"redhat-operators-xmnbf\" (UID: \"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a\") " pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.420359 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.739788 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.740138 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.820658 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 19:20:18 crc kubenswrapper[4828]: I1210 19:20:18.870488 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 19:20:19 crc kubenswrapper[4828]: I1210 19:20:19.119908 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 19:20:19 crc kubenswrapper[4828]: I1210 19:20:19.119968 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 19:20:19 crc kubenswrapper[4828]: I1210 19:20:19.339694 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 19:20:19 crc kubenswrapper[4828]: I1210 19:20:19.339745 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 19:20:19 crc kubenswrapper[4828]: I1210 19:20:19.373458 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 19:20:19 crc kubenswrapper[4828]: I1210 19:20:19.390055 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 19:20:20 crc kubenswrapper[4828]: I1210 19:20:20.128386 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 19:20:20 crc kubenswrapper[4828]: I1210 19:20:20.129037 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 19:20:20 crc kubenswrapper[4828]: W1210 19:20:20.631682 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddf32ed2_2731_497c_a021_6a00f495265f.slice/crio-149e691def5b33b1a1a967d48f2a38ebd954ff995beab863116acedb67250efc WatchSource:0}: Error finding container 149e691def5b33b1a1a967d48f2a38ebd954ff995beab863116acedb67250efc: Status 404 returned error can't find the container with id 149e691def5b33b1a1a967d48f2a38ebd954ff995beab863116acedb67250efc Dec 10 19:20:20 crc kubenswrapper[4828]: I1210 19:20:20.954190 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-56df8fb6b7-ptl7r" podUID="046c39f9-a9eb-45d4-9965-5575059fdcfc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.185:5353: i/o timeout" Dec 10 19:20:21 crc kubenswrapper[4828]: I1210 19:20:21.139414 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66756f49f6-r2lds" event={"ID":"ddf32ed2-2731-497c-a021-6a00f495265f","Type":"ContainerStarted","Data":"149e691def5b33b1a1a967d48f2a38ebd954ff995beab863116acedb67250efc"} Dec 10 19:20:21 crc kubenswrapper[4828]: I1210 19:20:21.140391 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" event={"ID":"ef7ea83b-99e3-4460-b3db-63b0fae43070","Type":"ContainerStarted","Data":"87da3edc14147c6aa655cd36b8561b454ac562daef1a885e756fe9087bdb62e6"} Dec 10 19:20:21 crc kubenswrapper[4828]: I1210 19:20:21.230229 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:20:21 crc kubenswrapper[4828]: I1210 19:20:21.230283 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:20:21 crc kubenswrapper[4828]: I1210 19:20:21.475132 4828 scope.go:117] "RemoveContainer" containerID="331a9ca0786faa74ad0eec3f00586a0ba2d92d7219a89fcc0d0fa25ae53c366e" Dec 10 19:20:21 crc kubenswrapper[4828]: I1210 19:20:21.642348 4828 scope.go:117] "RemoveContainer" containerID="75cfb358c76392b2d4e6c561fe78f4056bd3eb7f6a2498b6b261edd0fa58d233" Dec 10 19:20:22 crc kubenswrapper[4828]: I1210 19:20:22.061197 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xmnbf"] Dec 10 19:20:22 crc kubenswrapper[4828]: I1210 19:20:22.266897 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6598f5c59d-btf5k"] Dec 10 19:20:22 crc kubenswrapper[4828]: I1210 19:20:22.269277 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7cf7687b58-b784b" event={"ID":"339f67ce-2474-4399-a2ad-2e4bcfdd01af","Type":"ContainerStarted","Data":"27fe806867c74149f373bb796a2a0c33d073500e40134431fdca274c4afc9684"} Dec 10 19:20:22 crc kubenswrapper[4828]: I1210 19:20:22.270884 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:22 crc kubenswrapper[4828]: I1210 19:20:22.271179 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:22 crc kubenswrapper[4828]: I1210 19:20:22.306742 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7cf7687b58-b784b" podStartSLOduration=12.306722739 podStartE2EDuration="12.306722739s" podCreationTimestamp="2025-12-10 19:20:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:22.294432495 +0000 UTC m=+1502.805043500" watchObservedRunningTime="2025-12-10 19:20:22.306722739 +0000 UTC m=+1502.817333744" Dec 10 19:20:22 crc kubenswrapper[4828]: I1210 19:20:22.473515 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-56c96d9477-vps8l"] Dec 10 19:20:22 crc kubenswrapper[4828]: W1210 19:20:22.474876 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b65ecea_c14f_4702_be01_8952fabfb1be.slice/crio-14e7fd7fa150c73efb9aaa8a025c9b8cfb92f83ea7b5e899f1497a3fc51e5048 WatchSource:0}: Error finding container 14e7fd7fa150c73efb9aaa8a025c9b8cfb92f83ea7b5e899f1497a3fc51e5048: Status 404 returned error can't find the container with id 14e7fd7fa150c73efb9aaa8a025c9b8cfb92f83ea7b5e899f1497a3fc51e5048 Dec 10 19:20:23 crc kubenswrapper[4828]: I1210 19:20:23.283781 4828 generic.go:334] "Generic (PLEG): container finished" podID="ef7ea83b-99e3-4460-b3db-63b0fae43070" containerID="e15eddcbf82a51fcee4bed867488548b482aa968109b31d4dad7526bb32b85fa" exitCode=0 Dec 10 19:20:23 crc kubenswrapper[4828]: I1210 19:20:23.283946 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" event={"ID":"ef7ea83b-99e3-4460-b3db-63b0fae43070","Type":"ContainerDied","Data":"e15eddcbf82a51fcee4bed867488548b482aa968109b31d4dad7526bb32b85fa"} Dec 10 19:20:23 crc kubenswrapper[4828]: I1210 19:20:23.288413 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xmnbf" event={"ID":"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a","Type":"ContainerStarted","Data":"9a190a12c747d6f01a8236f0a84d43d5ea1b992eec960a0364ad5cea3401f1d4"} Dec 10 19:20:23 crc kubenswrapper[4828]: I1210 19:20:23.291834 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6598f5c59d-btf5k" event={"ID":"d425401e-33bc-4dc1-adce-44622ce940df","Type":"ContainerStarted","Data":"bc209ec438931bc6056471e347acea2c3d7c525eea8394e45f99680b3f07fd2c"} Dec 10 19:20:23 crc kubenswrapper[4828]: I1210 19:20:23.299690 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2n7h2" event={"ID":"acbabd11-f05a-4d18-a19a-6f1a74212a65","Type":"ContainerStarted","Data":"a416cfc1250e09a7ae12693c3a210bf1319994086d1e3b12d3e1032bd32df10d"} Dec 10 19:20:23 crc kubenswrapper[4828]: I1210 19:20:23.299716 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56c96d9477-vps8l" event={"ID":"2b65ecea-c14f-4702-be01-8952fabfb1be","Type":"ContainerStarted","Data":"14e7fd7fa150c73efb9aaa8a025c9b8cfb92f83ea7b5e899f1497a3fc51e5048"} Dec 10 19:20:23 crc kubenswrapper[4828]: I1210 19:20:23.310003 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-69884f6747-2l55b" event={"ID":"7ee9fcef-2371-4614-b67f-6cef46b20c6f","Type":"ContainerStarted","Data":"ee35b2917ea6414250655e6572528e37d807bec5c5fcd19822f87e5f57dd213f"} Dec 10 19:20:23 crc kubenswrapper[4828]: I1210 19:20:23.314768 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" event={"ID":"1bc1ebd1-1426-4def-a07d-5b0b561f2082","Type":"ContainerStarted","Data":"4ee9cf16cb130a5fc0ce21f29841e9dc12e41fe7e7da4dec2e2b251e80875b06"} Dec 10 19:20:23 crc kubenswrapper[4828]: I1210 19:20:23.317243 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66756f49f6-r2lds" event={"ID":"ddf32ed2-2731-497c-a021-6a00f495265f","Type":"ContainerStarted","Data":"8a66efac7a7e4f9bc7e860f04d479420bce2ca5ba3a1e8f3e01a2fa150047e11"} Dec 10 19:20:23 crc kubenswrapper[4828]: I1210 19:20:23.319774 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nkxxv" event={"ID":"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b","Type":"ContainerStarted","Data":"19fc8ad8f22009d9d47bef05d78283f3ad4397395f0f7469f014515fd84ccd08"} Dec 10 19:20:23 crc kubenswrapper[4828]: I1210 19:20:23.327846 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cdea527-5341-46d2-b11f-9c88c4b35469","Type":"ContainerStarted","Data":"c7528171774107299b6d02437a7b8bb95ec7146fca42ae05b221b1c8ec6c6844"} Dec 10 19:20:23 crc kubenswrapper[4828]: I1210 19:20:23.373647 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-2n7h2" podStartSLOduration=3.687441413 podStartE2EDuration="54.373615385s" podCreationTimestamp="2025-12-10 19:19:29 +0000 UTC" firstStartedPulling="2025-12-10 19:19:31.006002042 +0000 UTC m=+1451.516613047" lastFinishedPulling="2025-12-10 19:20:21.692176014 +0000 UTC m=+1502.202787019" observedRunningTime="2025-12-10 19:20:23.329517213 +0000 UTC m=+1503.840128218" watchObservedRunningTime="2025-12-10 19:20:23.373615385 +0000 UTC m=+1503.884226390" Dec 10 19:20:23 crc kubenswrapper[4828]: I1210 19:20:23.460246 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-nkxxv" podStartSLOduration=4.082718013 podStartE2EDuration="54.460225486s" podCreationTimestamp="2025-12-10 19:19:29 +0000 UTC" firstStartedPulling="2025-12-10 19:19:31.362965003 +0000 UTC m=+1451.873576008" lastFinishedPulling="2025-12-10 19:20:21.740472476 +0000 UTC m=+1502.251083481" observedRunningTime="2025-12-10 19:20:23.349458009 +0000 UTC m=+1503.860069034" watchObservedRunningTime="2025-12-10 19:20:23.460225486 +0000 UTC m=+1503.970836491" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.352170 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-sqwjh" event={"ID":"22554fda-f55d-4243-8416-7362ae946396","Type":"ContainerStarted","Data":"fb43fc1f48ae34006bf74d4d8397051d5eeda58f610595c35d22768fdbb8d948"} Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.363515 4828 generic.go:334] "Generic (PLEG): container finished" podID="5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" containerID="5f1f7213d7475e6ee288c2556ef8fb5123714ba9df2618833f23317f99b506e5" exitCode=0 Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.363622 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xmnbf" event={"ID":"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a","Type":"ContainerDied","Data":"5f1f7213d7475e6ee288c2556ef8fb5123714ba9df2618833f23317f99b506e5"} Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.372596 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-sqwjh" podStartSLOduration=4.490124497 podStartE2EDuration="54.372575803s" podCreationTimestamp="2025-12-10 19:19:30 +0000 UTC" firstStartedPulling="2025-12-10 19:19:31.901841624 +0000 UTC m=+1452.412452629" lastFinishedPulling="2025-12-10 19:20:21.78429292 +0000 UTC m=+1502.294903935" observedRunningTime="2025-12-10 19:20:24.368650199 +0000 UTC m=+1504.879261214" watchObservedRunningTime="2025-12-10 19:20:24.372575803 +0000 UTC m=+1504.883186808" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.374047 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-69884f6747-2l55b" event={"ID":"7ee9fcef-2371-4614-b67f-6cef46b20c6f","Type":"ContainerStarted","Data":"3f2628d638542a99cf54f894990b9501a9a0fc94edc53b6978988740c16d9d3d"} Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.380495 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" event={"ID":"1bc1ebd1-1426-4def-a07d-5b0b561f2082","Type":"ContainerStarted","Data":"59dc7c87be00c36f6592b27ab4f08321cc7c34735c94a13e6b1534938f21951e"} Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.394688 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6598f5c59d-btf5k" event={"ID":"d425401e-33bc-4dc1-adce-44622ce940df","Type":"ContainerStarted","Data":"09a8677322cd4d8e12156202d69769ed346ccbc33950dccfa9ae672ecacd7c5f"} Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.394749 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6598f5c59d-btf5k" event={"ID":"d425401e-33bc-4dc1-adce-44622ce940df","Type":"ContainerStarted","Data":"c77ec897b9cb0eb7c82e1075dcce4bd2253b1479b2e087a703378243d92117b3"} Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.395664 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.395726 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.401749 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66756f49f6-r2lds" event={"ID":"ddf32ed2-2731-497c-a021-6a00f495265f","Type":"ContainerStarted","Data":"300ec2101e4f20386ba3f3b3c21d97e458deb77fd9598c06b4391b95209e458c"} Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.407153 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.413272 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56c96d9477-vps8l" event={"ID":"2b65ecea-c14f-4702-be01-8952fabfb1be","Type":"ContainerStarted","Data":"4c9cb1a062c06f3644f51f776dbde61b4f7c7acf2b5f5e1f16f5c952b0c29c41"} Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.413325 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56c96d9477-vps8l" event={"ID":"2b65ecea-c14f-4702-be01-8952fabfb1be","Type":"ContainerStarted","Data":"e97fcef679d6622becf5191826473944376273d2046b2031d520961d0cada635"} Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.415518 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.426169 4828 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.426550 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-85965bf486-fgbxr" podStartSLOduration=5.453114918 podStartE2EDuration="14.426529884s" podCreationTimestamp="2025-12-10 19:20:10 +0000 UTC" firstStartedPulling="2025-12-10 19:20:12.518965367 +0000 UTC m=+1493.029576362" lastFinishedPulling="2025-12-10 19:20:21.492380323 +0000 UTC m=+1502.002991328" observedRunningTime="2025-12-10 19:20:24.402717376 +0000 UTC m=+1504.913328381" watchObservedRunningTime="2025-12-10 19:20:24.426529884 +0000 UTC m=+1504.937140889" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.427125 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" event={"ID":"ef7ea83b-99e3-4460-b3db-63b0fae43070","Type":"ContainerStarted","Data":"cb9f5250a01313095b6980a1c075687600d887952e18da51cd0f2dba40d40fb0"} Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.427702 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.429748 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-69884f6747-2l55b" podStartSLOduration=5.324855741 podStartE2EDuration="14.429714267s" podCreationTimestamp="2025-12-10 19:20:10 +0000 UTC" firstStartedPulling="2025-12-10 19:20:12.443272865 +0000 UTC m=+1492.953883870" lastFinishedPulling="2025-12-10 19:20:21.548131391 +0000 UTC m=+1502.058742396" observedRunningTime="2025-12-10 19:20:24.427401847 +0000 UTC m=+1504.938012852" watchObservedRunningTime="2025-12-10 19:20:24.429714267 +0000 UTC m=+1504.940325282" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.440597 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.440738 4828 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.465563 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.487627 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-66756f49f6-r2lds" podStartSLOduration=12.487597772 podStartE2EDuration="12.487597772s" podCreationTimestamp="2025-12-10 19:20:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:24.448324968 +0000 UTC m=+1504.958935973" watchObservedRunningTime="2025-12-10 19:20:24.487597772 +0000 UTC m=+1504.998208777" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.498370 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6598f5c59d-btf5k" podStartSLOduration=7.498347665 podStartE2EDuration="7.498347665s" podCreationTimestamp="2025-12-10 19:20:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:24.475620676 +0000 UTC m=+1504.986231681" watchObservedRunningTime="2025-12-10 19:20:24.498347665 +0000 UTC m=+1505.008958670" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.537651 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-56c96d9477-vps8l" podStartSLOduration=8.537628979 podStartE2EDuration="8.537628979s" podCreationTimestamp="2025-12-10 19:20:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:24.504118117 +0000 UTC m=+1505.014729132" watchObservedRunningTime="2025-12-10 19:20:24.537628979 +0000 UTC m=+1505.048239984" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.556073 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" podStartSLOduration=12.556047234 podStartE2EDuration="12.556047234s" podCreationTimestamp="2025-12-10 19:20:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:24.531401695 +0000 UTC m=+1505.042012700" watchObservedRunningTime="2025-12-10 19:20:24.556047234 +0000 UTC m=+1505.066658239" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.581865 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.581966 4828 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 19:20:24 crc kubenswrapper[4828]: I1210 19:20:24.588952 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 19:20:26 crc kubenswrapper[4828]: I1210 19:20:26.459253 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xmnbf" event={"ID":"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a","Type":"ContainerStarted","Data":"d1b567629ae86099880aff767da3052efee6a1b9fb6c3120df856e309ce2a1e5"} Dec 10 19:20:27 crc kubenswrapper[4828]: I1210 19:20:27.313089 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7cf7687b58-b784b" podUID="339f67ce-2474-4399-a2ad-2e4bcfdd01af" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.191:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 19:20:27 crc kubenswrapper[4828]: I1210 19:20:27.474052 4828 generic.go:334] "Generic (PLEG): container finished" podID="5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" containerID="d1b567629ae86099880aff767da3052efee6a1b9fb6c3120df856e309ce2a1e5" exitCode=0 Dec 10 19:20:27 crc kubenswrapper[4828]: I1210 19:20:27.474098 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xmnbf" event={"ID":"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a","Type":"ContainerDied","Data":"d1b567629ae86099880aff767da3052efee6a1b9fb6c3120df856e309ce2a1e5"} Dec 10 19:20:28 crc kubenswrapper[4828]: I1210 19:20:28.080202 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:28 crc kubenswrapper[4828]: I1210 19:20:28.292961 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:28 crc kubenswrapper[4828]: I1210 19:20:28.369300 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56798b757f-df8js"] Dec 10 19:20:28 crc kubenswrapper[4828]: I1210 19:20:28.369863 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56798b757f-df8js" podUID="95e1cafb-8bc1-4448-8738-4be3d7596e72" containerName="dnsmasq-dns" containerID="cri-o://de0625c431ca982a1b2ea712717546aa2c5ec199edbccddc32655bb808b04bda" gracePeriod=10 Dec 10 19:20:30 crc kubenswrapper[4828]: I1210 19:20:30.089543 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:30 crc kubenswrapper[4828]: I1210 19:20:30.232978 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:30 crc kubenswrapper[4828]: I1210 19:20:30.331880 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-56798b757f-df8js" podUID="95e1cafb-8bc1-4448-8738-4be3d7596e72" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.178:5353: connect: connection refused" Dec 10 19:20:30 crc kubenswrapper[4828]: I1210 19:20:30.569067 4828 generic.go:334] "Generic (PLEG): container finished" podID="95e1cafb-8bc1-4448-8738-4be3d7596e72" containerID="de0625c431ca982a1b2ea712717546aa2c5ec199edbccddc32655bb808b04bda" exitCode=0 Dec 10 19:20:30 crc kubenswrapper[4828]: I1210 19:20:30.569162 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56798b757f-df8js" event={"ID":"95e1cafb-8bc1-4448-8738-4be3d7596e72","Type":"ContainerDied","Data":"de0625c431ca982a1b2ea712717546aa2c5ec199edbccddc32655bb808b04bda"} Dec 10 19:20:30 crc kubenswrapper[4828]: I1210 19:20:30.581387 4828 generic.go:334] "Generic (PLEG): container finished" podID="d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b" containerID="19fc8ad8f22009d9d47bef05d78283f3ad4397395f0f7469f014515fd84ccd08" exitCode=0 Dec 10 19:20:30 crc kubenswrapper[4828]: I1210 19:20:30.581577 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nkxxv" event={"ID":"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b","Type":"ContainerDied","Data":"19fc8ad8f22009d9d47bef05d78283f3ad4397395f0f7469f014515fd84ccd08"} Dec 10 19:20:32 crc kubenswrapper[4828]: I1210 19:20:32.016366 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6598f5c59d-btf5k" Dec 10 19:20:32 crc kubenswrapper[4828]: I1210 19:20:32.095654 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7cf7687b58-b784b"] Dec 10 19:20:32 crc kubenswrapper[4828]: I1210 19:20:32.095925 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7cf7687b58-b784b" podUID="339f67ce-2474-4399-a2ad-2e4bcfdd01af" containerName="barbican-api-log" containerID="cri-o://fc3ea5b4555c800a8ca31a9c32426a49d13c9822a1a69181482217051a4cde86" gracePeriod=30 Dec 10 19:20:32 crc kubenswrapper[4828]: I1210 19:20:32.096021 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7cf7687b58-b784b" podUID="339f67ce-2474-4399-a2ad-2e4bcfdd01af" containerName="barbican-api" containerID="cri-o://27fe806867c74149f373bb796a2a0c33d073500e40134431fdca274c4afc9684" gracePeriod=30 Dec 10 19:20:32 crc kubenswrapper[4828]: I1210 19:20:32.630162 4828 generic.go:334] "Generic (PLEG): container finished" podID="339f67ce-2474-4399-a2ad-2e4bcfdd01af" containerID="fc3ea5b4555c800a8ca31a9c32426a49d13c9822a1a69181482217051a4cde86" exitCode=143 Dec 10 19:20:32 crc kubenswrapper[4828]: I1210 19:20:32.630382 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7cf7687b58-b784b" event={"ID":"339f67ce-2474-4399-a2ad-2e4bcfdd01af","Type":"ContainerDied","Data":"fc3ea5b4555c800a8ca31a9c32426a49d13c9822a1a69181482217051a4cde86"} Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.384056 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nkxxv" Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.468273 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rskbl\" (UniqueName: \"kubernetes.io/projected/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-kube-api-access-rskbl\") pod \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.468670 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-config-data\") pod \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.469004 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-combined-ca-bundle\") pod \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.469178 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-logs\") pod \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.469327 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-scripts\") pod \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\" (UID: \"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b\") " Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.469377 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-logs" (OuterVolumeSpecName: "logs") pod "d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b" (UID: "d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.470389 4828 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.475477 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-kube-api-access-rskbl" (OuterVolumeSpecName: "kube-api-access-rskbl") pod "d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b" (UID: "d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b"). InnerVolumeSpecName "kube-api-access-rskbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.476347 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-scripts" (OuterVolumeSpecName: "scripts") pod "d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b" (UID: "d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.507296 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b" (UID: "d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.508534 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-config-data" (OuterVolumeSpecName: "config-data") pod "d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b" (UID: "d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.572984 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.573025 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rskbl\" (UniqueName: \"kubernetes.io/projected/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-kube-api-access-rskbl\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.573039 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.573050 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.658098 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-nkxxv" event={"ID":"d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b","Type":"ContainerDied","Data":"c62774aab708c49f06edb6335d7fa2827ffaad464683a2d878d9e086e13703b1"} Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.658140 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c62774aab708c49f06edb6335d7fa2827ffaad464683a2d878d9e086e13703b1" Dec 10 19:20:34 crc kubenswrapper[4828]: I1210 19:20:34.658153 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-nkxxv" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.331746 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-56798b757f-df8js" podUID="95e1cafb-8bc1-4448-8738-4be3d7596e72" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.178:5353: connect: connection refused" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.528485 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-f568c66f6-nt8g5"] Dec 10 19:20:35 crc kubenswrapper[4828]: E1210 19:20:35.528993 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b" containerName="placement-db-sync" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.529006 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b" containerName="placement-db-sync" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.529207 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b" containerName="placement-db-sync" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.531598 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.534407 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.535685 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.535882 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.536023 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.536121 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-8gm2q" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.554258 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f568c66f6-nt8g5"] Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.592614 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-scripts\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.592684 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk864\" (UniqueName: \"kubernetes.io/projected/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-kube-api-access-tk864\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.592709 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-public-tls-certs\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.592731 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-logs\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.592763 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-config-data\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.592784 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-internal-tls-certs\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.593128 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-combined-ca-bundle\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.678744 4828 generic.go:334] "Generic (PLEG): container finished" podID="339f67ce-2474-4399-a2ad-2e4bcfdd01af" containerID="27fe806867c74149f373bb796a2a0c33d073500e40134431fdca274c4afc9684" exitCode=0 Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.678807 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7cf7687b58-b784b" event={"ID":"339f67ce-2474-4399-a2ad-2e4bcfdd01af","Type":"ContainerDied","Data":"27fe806867c74149f373bb796a2a0c33d073500e40134431fdca274c4afc9684"} Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.694984 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-scripts\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.695214 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk864\" (UniqueName: \"kubernetes.io/projected/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-kube-api-access-tk864\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.695246 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-public-tls-certs\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.695274 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-logs\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.695319 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-config-data\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.695341 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-internal-tls-certs\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.695524 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-combined-ca-bundle\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.699520 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-logs\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.704433 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-combined-ca-bundle\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.707305 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-public-tls-certs\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.709915 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-internal-tls-certs\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.715848 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-scripts\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.719097 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-config-data\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.719146 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk864\" (UniqueName: \"kubernetes.io/projected/06e3ba7a-1540-4093-b273-bc1c5ffa8ca6-kube-api-access-tk864\") pod \"placement-f568c66f6-nt8g5\" (UID: \"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6\") " pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:35 crc kubenswrapper[4828]: I1210 19:20:35.881368 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.497935 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.591404 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.615295 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-dns-svc\") pod \"95e1cafb-8bc1-4448-8738-4be3d7596e72\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.615355 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npkcf\" (UniqueName: \"kubernetes.io/projected/95e1cafb-8bc1-4448-8738-4be3d7596e72-kube-api-access-npkcf\") pod \"95e1cafb-8bc1-4448-8738-4be3d7596e72\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.615512 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-config-data\") pod \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.615563 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-ovsdbserver-nb\") pod \"95e1cafb-8bc1-4448-8738-4be3d7596e72\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.615592 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-config-data-custom\") pod \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.615640 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/339f67ce-2474-4399-a2ad-2e4bcfdd01af-logs\") pod \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.615683 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-config\") pod \"95e1cafb-8bc1-4448-8738-4be3d7596e72\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.615730 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-ovsdbserver-sb\") pod \"95e1cafb-8bc1-4448-8738-4be3d7596e72\" (UID: \"95e1cafb-8bc1-4448-8738-4be3d7596e72\") " Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.615771 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcxn9\" (UniqueName: \"kubernetes.io/projected/339f67ce-2474-4399-a2ad-2e4bcfdd01af-kube-api-access-fcxn9\") pod \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.615826 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-combined-ca-bundle\") pod \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\" (UID: \"339f67ce-2474-4399-a2ad-2e4bcfdd01af\") " Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.623295 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/339f67ce-2474-4399-a2ad-2e4bcfdd01af-logs" (OuterVolumeSpecName: "logs") pod "339f67ce-2474-4399-a2ad-2e4bcfdd01af" (UID: "339f67ce-2474-4399-a2ad-2e4bcfdd01af"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.665153 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95e1cafb-8bc1-4448-8738-4be3d7596e72-kube-api-access-npkcf" (OuterVolumeSpecName: "kube-api-access-npkcf") pod "95e1cafb-8bc1-4448-8738-4be3d7596e72" (UID: "95e1cafb-8bc1-4448-8738-4be3d7596e72"). InnerVolumeSpecName "kube-api-access-npkcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.665505 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "339f67ce-2474-4399-a2ad-2e4bcfdd01af" (UID: "339f67ce-2474-4399-a2ad-2e4bcfdd01af"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.666528 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/339f67ce-2474-4399-a2ad-2e4bcfdd01af-kube-api-access-fcxn9" (OuterVolumeSpecName: "kube-api-access-fcxn9") pod "339f67ce-2474-4399-a2ad-2e4bcfdd01af" (UID: "339f67ce-2474-4399-a2ad-2e4bcfdd01af"). InnerVolumeSpecName "kube-api-access-fcxn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.712603 4828 generic.go:334] "Generic (PLEG): container finished" podID="22554fda-f55d-4243-8416-7362ae946396" containerID="fb43fc1f48ae34006bf74d4d8397051d5eeda58f610595c35d22768fdbb8d948" exitCode=0 Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.712679 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-sqwjh" event={"ID":"22554fda-f55d-4243-8416-7362ae946396","Type":"ContainerDied","Data":"fb43fc1f48ae34006bf74d4d8397051d5eeda58f610595c35d22768fdbb8d948"} Dec 10 19:20:36 crc kubenswrapper[4828]: E1210 19:20:36.719736 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22554fda_f55d_4243_8416_7362ae946396.slice/crio-fb43fc1f48ae34006bf74d4d8397051d5eeda58f610595c35d22768fdbb8d948.scope\": RecentStats: unable to find data in memory cache]" Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.719763 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npkcf\" (UniqueName: \"kubernetes.io/projected/95e1cafb-8bc1-4448-8738-4be3d7596e72-kube-api-access-npkcf\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.719787 4828 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.719812 4828 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/339f67ce-2474-4399-a2ad-2e4bcfdd01af-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.719822 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcxn9\" (UniqueName: \"kubernetes.io/projected/339f67ce-2474-4399-a2ad-2e4bcfdd01af-kube-api-access-fcxn9\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.722321 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56798b757f-df8js" event={"ID":"95e1cafb-8bc1-4448-8738-4be3d7596e72","Type":"ContainerDied","Data":"2f16b3e08f526cf70a31cfec8160f100ddf4e5d63aadb36c081ce7216e2bd04a"} Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.722398 4828 scope.go:117] "RemoveContainer" containerID="de0625c431ca982a1b2ea712717546aa2c5ec199edbccddc32655bb808b04bda" Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.722348 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56798b757f-df8js" Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.727622 4828 generic.go:334] "Generic (PLEG): container finished" podID="acbabd11-f05a-4d18-a19a-6f1a74212a65" containerID="a416cfc1250e09a7ae12693c3a210bf1319994086d1e3b12d3e1032bd32df10d" exitCode=0 Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.727707 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2n7h2" event={"ID":"acbabd11-f05a-4d18-a19a-6f1a74212a65","Type":"ContainerDied","Data":"a416cfc1250e09a7ae12693c3a210bf1319994086d1e3b12d3e1032bd32df10d"} Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.731992 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7cf7687b58-b784b" event={"ID":"339f67ce-2474-4399-a2ad-2e4bcfdd01af","Type":"ContainerDied","Data":"991d0d54d2e48afe042e90163f9bc64719691578f5964443c3d76fd7f2d8382e"} Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.732080 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7cf7687b58-b784b" Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.778870 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f568c66f6-nt8g5"] Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.937968 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "339f67ce-2474-4399-a2ad-2e4bcfdd01af" (UID: "339f67ce-2474-4399-a2ad-2e4bcfdd01af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4828]: E1210 19:20:36.966838 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:800aa67a8f7a6cd334bd8ba44d1410e8aebc3aa6f347c5abc168de1222420526: Get \\\"https://quay.io/v2/podified-antelope-centos9/openstack-ceilometer-central/blobs/sha256:800aa67a8f7a6cd334bd8ba44d1410e8aebc3aa6f347c5abc168de1222420526\\\": context canceled\"" pod="openstack/ceilometer-0" podUID="9cdea527-5341-46d2-b11f-9c88c4b35469" Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.968667 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "95e1cafb-8bc1-4448-8738-4be3d7596e72" (UID: "95e1cafb-8bc1-4448-8738-4be3d7596e72"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:36 crc kubenswrapper[4828]: I1210 19:20:36.972351 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "95e1cafb-8bc1-4448-8738-4be3d7596e72" (UID: "95e1cafb-8bc1-4448-8738-4be3d7596e72"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.003423 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "95e1cafb-8bc1-4448-8738-4be3d7596e72" (UID: "95e1cafb-8bc1-4448-8738-4be3d7596e72"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.025564 4828 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.025596 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.025608 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.025620 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.060852 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-config" (OuterVolumeSpecName: "config") pod "95e1cafb-8bc1-4448-8738-4be3d7596e72" (UID: "95e1cafb-8bc1-4448-8738-4be3d7596e72"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.081758 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-config-data" (OuterVolumeSpecName: "config-data") pod "339f67ce-2474-4399-a2ad-2e4bcfdd01af" (UID: "339f67ce-2474-4399-a2ad-2e4bcfdd01af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.115012 4828 scope.go:117] "RemoveContainer" containerID="00e7c5f0ecf6fb8966aeaf550150cb81d215d281884ca137761f9b32e0555d99" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.127505 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/339f67ce-2474-4399-a2ad-2e4bcfdd01af-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.127541 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95e1cafb-8bc1-4448-8738-4be3d7596e72-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.257303 4828 scope.go:117] "RemoveContainer" containerID="27fe806867c74149f373bb796a2a0c33d073500e40134431fdca274c4afc9684" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.288108 4828 scope.go:117] "RemoveContainer" containerID="fc3ea5b4555c800a8ca31a9c32426a49d13c9822a1a69181482217051a4cde86" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.419139 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7cf7687b58-b784b"] Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.429266 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7cf7687b58-b784b"] Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.448841 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56798b757f-df8js"] Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.460486 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56798b757f-df8js"] Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.745052 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xmnbf" event={"ID":"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a","Type":"ContainerStarted","Data":"aba92744fea3340582f4b08edcdecf604363ec39c5ed811390e34f2ec576d930"} Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.748434 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f568c66f6-nt8g5" event={"ID":"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6","Type":"ContainerStarted","Data":"56bc73ad28896ce7224e95afc579ebd73410d3793bfd940667ab44f190d486c2"} Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.748550 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f568c66f6-nt8g5" event={"ID":"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6","Type":"ContainerStarted","Data":"6c027b53485e6db70fb88433b6520c5e4306bafdb8870ad1ab6cdb28055465b5"} Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.748656 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.748725 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.748792 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f568c66f6-nt8g5" event={"ID":"06e3ba7a-1540-4093-b273-bc1c5ffa8ca6","Type":"ContainerStarted","Data":"cf19b997947986ae1b532ae64beb4621ed120d899037476fd9499f6284bde239"} Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.750569 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cdea527-5341-46d2-b11f-9c88c4b35469","Type":"ContainerStarted","Data":"5b5a45acf1ff1b285a969bc0e3e24c5219eba7c41097a425a5d9b807f4b7147d"} Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.750788 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9cdea527-5341-46d2-b11f-9c88c4b35469" containerName="ceilometer-notification-agent" containerID="cri-o://c00655fab2f22d8330b86cf36e17d7861f6fd58142cbd516fb930de7f29f0494" gracePeriod=30 Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.751148 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.751252 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9cdea527-5341-46d2-b11f-9c88c4b35469" containerName="proxy-httpd" containerID="cri-o://5b5a45acf1ff1b285a969bc0e3e24c5219eba7c41097a425a5d9b807f4b7147d" gracePeriod=30 Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.751350 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9cdea527-5341-46d2-b11f-9c88c4b35469" containerName="sg-core" containerID="cri-o://c7528171774107299b6d02437a7b8bb95ec7146fca42ae05b221b1c8ec6c6844" gracePeriod=30 Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.790667 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xmnbf" podStartSLOduration=16.012208811 podStartE2EDuration="19.790651957s" podCreationTimestamp="2025-12-10 19:20:18 +0000 UTC" firstStartedPulling="2025-12-10 19:20:24.369981174 +0000 UTC m=+1504.880592169" lastFinishedPulling="2025-12-10 19:20:28.14842432 +0000 UTC m=+1508.659035315" observedRunningTime="2025-12-10 19:20:37.78465886 +0000 UTC m=+1518.295269865" watchObservedRunningTime="2025-12-10 19:20:37.790651957 +0000 UTC m=+1518.301262962" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.808699 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="339f67ce-2474-4399-a2ad-2e4bcfdd01af" path="/var/lib/kubelet/pods/339f67ce-2474-4399-a2ad-2e4bcfdd01af/volumes" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.809659 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95e1cafb-8bc1-4448-8738-4be3d7596e72" path="/var/lib/kubelet/pods/95e1cafb-8bc1-4448-8738-4be3d7596e72/volumes" Dec 10 19:20:37 crc kubenswrapper[4828]: I1210 19:20:37.862688 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-f568c66f6-nt8g5" podStartSLOduration=2.862667314 podStartE2EDuration="2.862667314s" podCreationTimestamp="2025-12-10 19:20:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:37.844240529 +0000 UTC m=+1518.354851534" watchObservedRunningTime="2025-12-10 19:20:37.862667314 +0000 UTC m=+1518.373278319" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.149875 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.232955 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2n7h2" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.257586 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-db-sync-config-data\") pod \"22554fda-f55d-4243-8416-7362ae946396\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.257637 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-scripts\") pod \"22554fda-f55d-4243-8416-7362ae946396\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.257747 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-config-data\") pod \"22554fda-f55d-4243-8416-7362ae946396\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.257792 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlh25\" (UniqueName: \"kubernetes.io/projected/22554fda-f55d-4243-8416-7362ae946396-kube-api-access-nlh25\") pod \"22554fda-f55d-4243-8416-7362ae946396\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.257890 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/22554fda-f55d-4243-8416-7362ae946396-etc-machine-id\") pod \"22554fda-f55d-4243-8416-7362ae946396\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.258066 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-combined-ca-bundle\") pod \"22554fda-f55d-4243-8416-7362ae946396\" (UID: \"22554fda-f55d-4243-8416-7362ae946396\") " Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.261932 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/22554fda-f55d-4243-8416-7362ae946396-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "22554fda-f55d-4243-8416-7362ae946396" (UID: "22554fda-f55d-4243-8416-7362ae946396"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.265565 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-scripts" (OuterVolumeSpecName: "scripts") pod "22554fda-f55d-4243-8416-7362ae946396" (UID: "22554fda-f55d-4243-8416-7362ae946396"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.266028 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22554fda-f55d-4243-8416-7362ae946396-kube-api-access-nlh25" (OuterVolumeSpecName: "kube-api-access-nlh25") pod "22554fda-f55d-4243-8416-7362ae946396" (UID: "22554fda-f55d-4243-8416-7362ae946396"). InnerVolumeSpecName "kube-api-access-nlh25". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.266074 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "22554fda-f55d-4243-8416-7362ae946396" (UID: "22554fda-f55d-4243-8416-7362ae946396"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.298132 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "22554fda-f55d-4243-8416-7362ae946396" (UID: "22554fda-f55d-4243-8416-7362ae946396"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.329397 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-config-data" (OuterVolumeSpecName: "config-data") pod "22554fda-f55d-4243-8416-7362ae946396" (UID: "22554fda-f55d-4243-8416-7362ae946396"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.361409 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acbabd11-f05a-4d18-a19a-6f1a74212a65-config-data\") pod \"acbabd11-f05a-4d18-a19a-6f1a74212a65\" (UID: \"acbabd11-f05a-4d18-a19a-6f1a74212a65\") " Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.361657 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acbabd11-f05a-4d18-a19a-6f1a74212a65-combined-ca-bundle\") pod \"acbabd11-f05a-4d18-a19a-6f1a74212a65\" (UID: \"acbabd11-f05a-4d18-a19a-6f1a74212a65\") " Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.361686 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrz7q\" (UniqueName: \"kubernetes.io/projected/acbabd11-f05a-4d18-a19a-6f1a74212a65-kube-api-access-qrz7q\") pod \"acbabd11-f05a-4d18-a19a-6f1a74212a65\" (UID: \"acbabd11-f05a-4d18-a19a-6f1a74212a65\") " Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.362187 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.362209 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.362222 4828 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.362231 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22554fda-f55d-4243-8416-7362ae946396-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.362240 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlh25\" (UniqueName: \"kubernetes.io/projected/22554fda-f55d-4243-8416-7362ae946396-kube-api-access-nlh25\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.362251 4828 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/22554fda-f55d-4243-8416-7362ae946396-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.365639 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acbabd11-f05a-4d18-a19a-6f1a74212a65-kube-api-access-qrz7q" (OuterVolumeSpecName: "kube-api-access-qrz7q") pod "acbabd11-f05a-4d18-a19a-6f1a74212a65" (UID: "acbabd11-f05a-4d18-a19a-6f1a74212a65"). InnerVolumeSpecName "kube-api-access-qrz7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.405304 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acbabd11-f05a-4d18-a19a-6f1a74212a65-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "acbabd11-f05a-4d18-a19a-6f1a74212a65" (UID: "acbabd11-f05a-4d18-a19a-6f1a74212a65"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.421640 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.421679 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.450404 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acbabd11-f05a-4d18-a19a-6f1a74212a65-config-data" (OuterVolumeSpecName: "config-data") pod "acbabd11-f05a-4d18-a19a-6f1a74212a65" (UID: "acbabd11-f05a-4d18-a19a-6f1a74212a65"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.464650 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acbabd11-f05a-4d18-a19a-6f1a74212a65-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.464696 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrz7q\" (UniqueName: \"kubernetes.io/projected/acbabd11-f05a-4d18-a19a-6f1a74212a65-kube-api-access-qrz7q\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.464709 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/acbabd11-f05a-4d18-a19a-6f1a74212a65-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.772301 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-sqwjh" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.772335 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-sqwjh" event={"ID":"22554fda-f55d-4243-8416-7362ae946396","Type":"ContainerDied","Data":"2d382ab9f15f453a6e87939a81e2d1205c250eb7ce031cdd0be141bea4a0a61d"} Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.773687 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d382ab9f15f453a6e87939a81e2d1205c250eb7ce031cdd0be141bea4a0a61d" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.778139 4828 generic.go:334] "Generic (PLEG): container finished" podID="9cdea527-5341-46d2-b11f-9c88c4b35469" containerID="5b5a45acf1ff1b285a969bc0e3e24c5219eba7c41097a425a5d9b807f4b7147d" exitCode=0 Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.778180 4828 generic.go:334] "Generic (PLEG): container finished" podID="9cdea527-5341-46d2-b11f-9c88c4b35469" containerID="c7528171774107299b6d02437a7b8bb95ec7146fca42ae05b221b1c8ec6c6844" exitCode=2 Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.778231 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cdea527-5341-46d2-b11f-9c88c4b35469","Type":"ContainerDied","Data":"5b5a45acf1ff1b285a969bc0e3e24c5219eba7c41097a425a5d9b807f4b7147d"} Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.778262 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cdea527-5341-46d2-b11f-9c88c4b35469","Type":"ContainerDied","Data":"c7528171774107299b6d02437a7b8bb95ec7146fca42ae05b221b1c8ec6c6844"} Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.783166 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2n7h2" Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.783860 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2n7h2" event={"ID":"acbabd11-f05a-4d18-a19a-6f1a74212a65","Type":"ContainerDied","Data":"d21113b0f8bcddebee990f72dd664ff1bce1eba5fb5efd221d5984e4a358dc9f"} Dec 10 19:20:38 crc kubenswrapper[4828]: I1210 19:20:38.783912 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d21113b0f8bcddebee990f72dd664ff1bce1eba5fb5efd221d5984e4a358dc9f" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.063860 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-s5bns"] Dec 10 19:20:39 crc kubenswrapper[4828]: E1210 19:20:39.064679 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="339f67ce-2474-4399-a2ad-2e4bcfdd01af" containerName="barbican-api" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.064699 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="339f67ce-2474-4399-a2ad-2e4bcfdd01af" containerName="barbican-api" Dec 10 19:20:39 crc kubenswrapper[4828]: E1210 19:20:39.064715 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22554fda-f55d-4243-8416-7362ae946396" containerName="cinder-db-sync" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.064722 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="22554fda-f55d-4243-8416-7362ae946396" containerName="cinder-db-sync" Dec 10 19:20:39 crc kubenswrapper[4828]: E1210 19:20:39.064737 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="339f67ce-2474-4399-a2ad-2e4bcfdd01af" containerName="barbican-api-log" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.064744 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="339f67ce-2474-4399-a2ad-2e4bcfdd01af" containerName="barbican-api-log" Dec 10 19:20:39 crc kubenswrapper[4828]: E1210 19:20:39.064758 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95e1cafb-8bc1-4448-8738-4be3d7596e72" containerName="init" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.064763 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="95e1cafb-8bc1-4448-8738-4be3d7596e72" containerName="init" Dec 10 19:20:39 crc kubenswrapper[4828]: E1210 19:20:39.064774 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95e1cafb-8bc1-4448-8738-4be3d7596e72" containerName="dnsmasq-dns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.064780 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="95e1cafb-8bc1-4448-8738-4be3d7596e72" containerName="dnsmasq-dns" Dec 10 19:20:39 crc kubenswrapper[4828]: E1210 19:20:39.064818 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acbabd11-f05a-4d18-a19a-6f1a74212a65" containerName="heat-db-sync" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.064826 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="acbabd11-f05a-4d18-a19a-6f1a74212a65" containerName="heat-db-sync" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.065070 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="339f67ce-2474-4399-a2ad-2e4bcfdd01af" containerName="barbican-api" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.065085 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="acbabd11-f05a-4d18-a19a-6f1a74212a65" containerName="heat-db-sync" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.065095 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="95e1cafb-8bc1-4448-8738-4be3d7596e72" containerName="dnsmasq-dns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.065111 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="22554fda-f55d-4243-8416-7362ae946396" containerName="cinder-db-sync" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.065120 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="339f67ce-2474-4399-a2ad-2e4bcfdd01af" containerName="barbican-api-log" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.076332 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.207598 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.207696 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb9hw\" (UniqueName: \"kubernetes.io/projected/fd2aa65d-1eae-421a-9198-684337484b92-kube-api-access-lb9hw\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.207762 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-config\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.208092 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.208276 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-dns-svc\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.208349 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.211335 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.224769 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-s5bns"] Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.224891 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.232178 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wbmsh" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.232543 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.232814 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.233110 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.242047 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.316541 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb9hw\" (UniqueName: \"kubernetes.io/projected/fd2aa65d-1eae-421a-9198-684337484b92-kube-api-access-lb9hw\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.316605 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-config\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.316634 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj652\" (UniqueName: \"kubernetes.io/projected/c5297488-c2cd-4c22-906b-2df48517ff45-kube-api-access-kj652\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.316663 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-scripts\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.316714 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.316783 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.316828 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-dns-svc\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.316860 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.316880 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.316922 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5297488-c2cd-4c22-906b-2df48517ff45-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.316959 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-config-data\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.316984 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.317866 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.318712 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-config\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.333488 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.339461 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-dns-svc\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.340589 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.369694 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb9hw\" (UniqueName: \"kubernetes.io/projected/fd2aa65d-1eae-421a-9198-684337484b92-kube-api-access-lb9hw\") pod \"dnsmasq-dns-6578955fd5-s5bns\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.373811 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.376450 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.389421 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.412439 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.419127 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.419216 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.419282 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5297488-c2cd-4c22-906b-2df48517ff45-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.419337 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-config-data\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.419410 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj652\" (UniqueName: \"kubernetes.io/projected/c5297488-c2cd-4c22-906b-2df48517ff45-kube-api-access-kj652\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.419444 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-scripts\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.419658 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5297488-c2cd-4c22-906b-2df48517ff45-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.437177 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.444387 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-scripts\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.446064 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.463404 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.468201 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-config-data\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.481700 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj652\" (UniqueName: \"kubernetes.io/projected/c5297488-c2cd-4c22-906b-2df48517ff45-kube-api-access-kj652\") pod \"cinder-scheduler-0\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.484645 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xmnbf" podUID="5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" containerName="registry-server" probeResult="failure" output=< Dec 10 19:20:39 crc kubenswrapper[4828]: timeout: failed to connect service ":50051" within 1s Dec 10 19:20:39 crc kubenswrapper[4828]: > Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.521538 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.521629 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-logs\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.521711 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-config-data-custom\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.521761 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptvb6\" (UniqueName: \"kubernetes.io/projected/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-kube-api-access-ptvb6\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.522432 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-scripts\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.522582 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-config-data\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.522723 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.543463 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.627119 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.627197 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-logs\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.627307 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-config-data-custom\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.627587 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptvb6\" (UniqueName: \"kubernetes.io/projected/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-kube-api-access-ptvb6\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.627632 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-scripts\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.627816 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-config-data\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.627921 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.628110 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.629240 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-logs\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.633116 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.639404 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-config-data-custom\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.639850 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-config-data\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.639973 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-scripts\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.670822 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptvb6\" (UniqueName: \"kubernetes.io/projected/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-kube-api-access-ptvb6\") pod \"cinder-api-0\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " pod="openstack/cinder-api-0" Dec 10 19:20:39 crc kubenswrapper[4828]: I1210 19:20:39.708668 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 19:20:40 crc kubenswrapper[4828]: I1210 19:20:40.288164 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:20:40 crc kubenswrapper[4828]: I1210 19:20:40.335677 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-s5bns"] Dec 10 19:20:40 crc kubenswrapper[4828]: W1210 19:20:40.341151 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd2aa65d_1eae_421a_9198_684337484b92.slice/crio-bae93beb40f5418760d4f13a83c452e9fb4a37adf58d5674257fae52b0ea95a0 WatchSource:0}: Error finding container bae93beb40f5418760d4f13a83c452e9fb4a37adf58d5674257fae52b0ea95a0: Status 404 returned error can't find the container with id bae93beb40f5418760d4f13a83c452e9fb4a37adf58d5674257fae52b0ea95a0 Dec 10 19:20:40 crc kubenswrapper[4828]: I1210 19:20:40.534769 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:20:40 crc kubenswrapper[4828]: I1210 19:20:40.830517 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c5297488-c2cd-4c22-906b-2df48517ff45","Type":"ContainerStarted","Data":"cb51f1d75749de9a2640714f4b5c7523910d452007200eb5fd3d66f70bb4f070"} Dec 10 19:20:40 crc kubenswrapper[4828]: I1210 19:20:40.841078 4828 generic.go:334] "Generic (PLEG): container finished" podID="9cdea527-5341-46d2-b11f-9c88c4b35469" containerID="c00655fab2f22d8330b86cf36e17d7861f6fd58142cbd516fb930de7f29f0494" exitCode=0 Dec 10 19:20:40 crc kubenswrapper[4828]: I1210 19:20:40.841159 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cdea527-5341-46d2-b11f-9c88c4b35469","Type":"ContainerDied","Data":"c00655fab2f22d8330b86cf36e17d7861f6fd58142cbd516fb930de7f29f0494"} Dec 10 19:20:40 crc kubenswrapper[4828]: I1210 19:20:40.848962 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-s5bns" event={"ID":"fd2aa65d-1eae-421a-9198-684337484b92","Type":"ContainerStarted","Data":"3f279e3ec9b46a37bad62bca048f38da14b00d3ddfe4cf78080f569f2e589218"} Dec 10 19:20:40 crc kubenswrapper[4828]: I1210 19:20:40.849001 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-s5bns" event={"ID":"fd2aa65d-1eae-421a-9198-684337484b92","Type":"ContainerStarted","Data":"bae93beb40f5418760d4f13a83c452e9fb4a37adf58d5674257fae52b0ea95a0"} Dec 10 19:20:40 crc kubenswrapper[4828]: I1210 19:20:40.856582 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8cb3c5d9-61fb-415a-a644-75803d3f7c8f","Type":"ContainerStarted","Data":"dc140dccad70f788c8579c38684802e86b64741f2ccc404dfd77524bfc74d057"} Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.299964 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.341641 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7cf7687b58-b784b" podUID="339f67ce-2474-4399-a2ad-2e4bcfdd01af" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.191:9311/healthcheck\": dial tcp 10.217.0.191:9311: i/o timeout (Client.Timeout exceeded while awaiting headers)" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.343052 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7cf7687b58-b784b" podUID="339f67ce-2474-4399-a2ad-2e4bcfdd01af" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.191:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.382374 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-config-data\") pod \"9cdea527-5341-46d2-b11f-9c88c4b35469\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.382540 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-scripts\") pod \"9cdea527-5341-46d2-b11f-9c88c4b35469\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.382591 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cdea527-5341-46d2-b11f-9c88c4b35469-log-httpd\") pod \"9cdea527-5341-46d2-b11f-9c88c4b35469\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.382681 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-sg-core-conf-yaml\") pod \"9cdea527-5341-46d2-b11f-9c88c4b35469\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.382741 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-combined-ca-bundle\") pod \"9cdea527-5341-46d2-b11f-9c88c4b35469\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.382771 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dntxb\" (UniqueName: \"kubernetes.io/projected/9cdea527-5341-46d2-b11f-9c88c4b35469-kube-api-access-dntxb\") pod \"9cdea527-5341-46d2-b11f-9c88c4b35469\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.382842 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cdea527-5341-46d2-b11f-9c88c4b35469-run-httpd\") pod \"9cdea527-5341-46d2-b11f-9c88c4b35469\" (UID: \"9cdea527-5341-46d2-b11f-9c88c4b35469\") " Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.383633 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cdea527-5341-46d2-b11f-9c88c4b35469-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9cdea527-5341-46d2-b11f-9c88c4b35469" (UID: "9cdea527-5341-46d2-b11f-9c88c4b35469"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.390074 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cdea527-5341-46d2-b11f-9c88c4b35469-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9cdea527-5341-46d2-b11f-9c88c4b35469" (UID: "9cdea527-5341-46d2-b11f-9c88c4b35469"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.419350 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-scripts" (OuterVolumeSpecName: "scripts") pod "9cdea527-5341-46d2-b11f-9c88c4b35469" (UID: "9cdea527-5341-46d2-b11f-9c88c4b35469"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.419353 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cdea527-5341-46d2-b11f-9c88c4b35469-kube-api-access-dntxb" (OuterVolumeSpecName: "kube-api-access-dntxb") pod "9cdea527-5341-46d2-b11f-9c88c4b35469" (UID: "9cdea527-5341-46d2-b11f-9c88c4b35469"). InnerVolumeSpecName "kube-api-access-dntxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.449668 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9cdea527-5341-46d2-b11f-9c88c4b35469" (UID: "9cdea527-5341-46d2-b11f-9c88c4b35469"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.485549 4828 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cdea527-5341-46d2-b11f-9c88c4b35469-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.485586 4828 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.485598 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dntxb\" (UniqueName: \"kubernetes.io/projected/9cdea527-5341-46d2-b11f-9c88c4b35469-kube-api-access-dntxb\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.485606 4828 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9cdea527-5341-46d2-b11f-9c88c4b35469-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.485614 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.523947 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-config-data" (OuterVolumeSpecName: "config-data") pod "9cdea527-5341-46d2-b11f-9c88c4b35469" (UID: "9cdea527-5341-46d2-b11f-9c88c4b35469"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.581444 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9cdea527-5341-46d2-b11f-9c88c4b35469" (UID: "9cdea527-5341-46d2-b11f-9c88c4b35469"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.589244 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.589290 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cdea527-5341-46d2-b11f-9c88c4b35469-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.932097 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9cdea527-5341-46d2-b11f-9c88c4b35469","Type":"ContainerDied","Data":"2beb37740c7ec969b0b1259c125767b59cabcff63fcdc5935204f3c283007fbd"} Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.932157 4828 scope.go:117] "RemoveContainer" containerID="5b5a45acf1ff1b285a969bc0e3e24c5219eba7c41097a425a5d9b807f4b7147d" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.932360 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.947891 4828 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a] : Timed out while waiting for systemd to remove kubepods-besteffort-pod1f2eb8fc_198b_4cf5_bcc9_e47f35b6324a.slice" Dec 10 19:20:41 crc kubenswrapper[4828]: E1210 19:20:41.947941 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a] : Timed out while waiting for systemd to remove kubepods-besteffort-pod1f2eb8fc_198b_4cf5_bcc9_e47f35b6324a.slice" pod="openstack/keystone-bootstrap-gwbxv" podUID="1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a" Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.962147 4828 generic.go:334] "Generic (PLEG): container finished" podID="fd2aa65d-1eae-421a-9198-684337484b92" containerID="3f279e3ec9b46a37bad62bca048f38da14b00d3ddfe4cf78080f569f2e589218" exitCode=0 Dec 10 19:20:41 crc kubenswrapper[4828]: I1210 19:20:41.962517 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-s5bns" event={"ID":"fd2aa65d-1eae-421a-9198-684337484b92","Type":"ContainerDied","Data":"3f279e3ec9b46a37bad62bca048f38da14b00d3ddfe4cf78080f569f2e589218"} Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.003922 4828 scope.go:117] "RemoveContainer" containerID="c7528171774107299b6d02437a7b8bb95ec7146fca42ae05b221b1c8ec6c6844" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.053508 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.054051 4828 scope.go:117] "RemoveContainer" containerID="c00655fab2f22d8330b86cf36e17d7861f6fd58142cbd516fb930de7f29f0494" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.091901 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.170664 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:20:42 crc kubenswrapper[4828]: E1210 19:20:42.171360 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cdea527-5341-46d2-b11f-9c88c4b35469" containerName="ceilometer-notification-agent" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.171379 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cdea527-5341-46d2-b11f-9c88c4b35469" containerName="ceilometer-notification-agent" Dec 10 19:20:42 crc kubenswrapper[4828]: E1210 19:20:42.171400 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cdea527-5341-46d2-b11f-9c88c4b35469" containerName="proxy-httpd" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.171408 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cdea527-5341-46d2-b11f-9c88c4b35469" containerName="proxy-httpd" Dec 10 19:20:42 crc kubenswrapper[4828]: E1210 19:20:42.171456 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cdea527-5341-46d2-b11f-9c88c4b35469" containerName="sg-core" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.171466 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cdea527-5341-46d2-b11f-9c88c4b35469" containerName="sg-core" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.171773 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cdea527-5341-46d2-b11f-9c88c4b35469" containerName="sg-core" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.171828 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cdea527-5341-46d2-b11f-9c88c4b35469" containerName="ceilometer-notification-agent" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.171884 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cdea527-5341-46d2-b11f-9c88c4b35469" containerName="proxy-httpd" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.199698 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.199896 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.214646 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.214918 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.345955 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.347099 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-scripts\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.347177 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01a17451-e170-42ad-9082-d91eb11e237a-run-httpd\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.347270 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6hb7\" (UniqueName: \"kubernetes.io/projected/01a17451-e170-42ad-9082-d91eb11e237a-kube-api-access-k6hb7\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.347565 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-config-data\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.347646 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01a17451-e170-42ad-9082-d91eb11e237a-log-httpd\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.347720 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.449592 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01a17451-e170-42ad-9082-d91eb11e237a-run-httpd\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.449740 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-scripts\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.449863 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6hb7\" (UniqueName: \"kubernetes.io/projected/01a17451-e170-42ad-9082-d91eb11e237a-kube-api-access-k6hb7\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.450117 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01a17451-e170-42ad-9082-d91eb11e237a-run-httpd\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.450256 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-config-data\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.450358 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01a17451-e170-42ad-9082-d91eb11e237a-log-httpd\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.450457 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.450590 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.451179 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01a17451-e170-42ad-9082-d91eb11e237a-log-httpd\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.455511 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.462666 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-config-data\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.475984 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.477879 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6hb7\" (UniqueName: \"kubernetes.io/projected/01a17451-e170-42ad-9082-d91eb11e237a-kube-api-access-k6hb7\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.479512 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-scripts\") pod \"ceilometer-0\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " pod="openstack/ceilometer-0" Dec 10 19:20:42 crc kubenswrapper[4828]: I1210 19:20:42.542372 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:20:43 crc kubenswrapper[4828]: I1210 19:20:43.003212 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8cb3c5d9-61fb-415a-a644-75803d3f7c8f","Type":"ContainerStarted","Data":"00139c641c2eb6e20a80832cd0c631ec70582d91abe9adaad0315426b3e6248a"} Dec 10 19:20:43 crc kubenswrapper[4828]: I1210 19:20:43.007073 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c5297488-c2cd-4c22-906b-2df48517ff45","Type":"ContainerStarted","Data":"99f10991c34d4247de761b42a0409fa7fb664ea2b7f9bde6181574e2257aa8b7"} Dec 10 19:20:43 crc kubenswrapper[4828]: I1210 19:20:43.016056 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-gwbxv" Dec 10 19:20:43 crc kubenswrapper[4828]: I1210 19:20:43.018107 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-s5bns" event={"ID":"fd2aa65d-1eae-421a-9198-684337484b92","Type":"ContainerStarted","Data":"eba8b69562e3c40b6f7239cf498dc8bc63213a06fbf4e6be415111e00411e688"} Dec 10 19:20:43 crc kubenswrapper[4828]: I1210 19:20:43.018224 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:43 crc kubenswrapper[4828]: I1210 19:20:43.059735 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:20:43 crc kubenswrapper[4828]: I1210 19:20:43.063329 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-s5bns" podStartSLOduration=4.063311514 podStartE2EDuration="4.063311514s" podCreationTimestamp="2025-12-10 19:20:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:43.050345932 +0000 UTC m=+1523.560956957" watchObservedRunningTime="2025-12-10 19:20:43.063311514 +0000 UTC m=+1523.573922519" Dec 10 19:20:43 crc kubenswrapper[4828]: I1210 19:20:43.257364 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:20:43 crc kubenswrapper[4828]: I1210 19:20:43.332736 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:43 crc kubenswrapper[4828]: I1210 19:20:43.803670 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cdea527-5341-46d2-b11f-9c88c4b35469" path="/var/lib/kubelet/pods/9cdea527-5341-46d2-b11f-9c88c4b35469/volumes" Dec 10 19:20:44 crc kubenswrapper[4828]: I1210 19:20:44.035898 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c5297488-c2cd-4c22-906b-2df48517ff45","Type":"ContainerStarted","Data":"173bd8216ca59a312cafa05f0f282643d70e91472d040004aae16edd4c800577"} Dec 10 19:20:44 crc kubenswrapper[4828]: I1210 19:20:44.042149 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01a17451-e170-42ad-9082-d91eb11e237a","Type":"ContainerStarted","Data":"f37198571727cf7744f0fee52aa280d9a195a7dc1d503360bf2ed01ef2567f53"} Dec 10 19:20:44 crc kubenswrapper[4828]: I1210 19:20:44.046829 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="8cb3c5d9-61fb-415a-a644-75803d3f7c8f" containerName="cinder-api-log" containerID="cri-o://00139c641c2eb6e20a80832cd0c631ec70582d91abe9adaad0315426b3e6248a" gracePeriod=30 Dec 10 19:20:44 crc kubenswrapper[4828]: I1210 19:20:44.046996 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8cb3c5d9-61fb-415a-a644-75803d3f7c8f","Type":"ContainerStarted","Data":"98e81aab581da45b5907a2bcaf76c5664a8e065cd952317d5fb3991b83648371"} Dec 10 19:20:44 crc kubenswrapper[4828]: I1210 19:20:44.047074 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 10 19:20:44 crc kubenswrapper[4828]: I1210 19:20:44.047108 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="8cb3c5d9-61fb-415a-a644-75803d3f7c8f" containerName="cinder-api" containerID="cri-o://98e81aab581da45b5907a2bcaf76c5664a8e065cd952317d5fb3991b83648371" gracePeriod=30 Dec 10 19:20:44 crc kubenswrapper[4828]: I1210 19:20:44.065823 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.023132346 podStartE2EDuration="5.065791644s" podCreationTimestamp="2025-12-10 19:20:39 +0000 UTC" firstStartedPulling="2025-12-10 19:20:40.302880388 +0000 UTC m=+1520.813491393" lastFinishedPulling="2025-12-10 19:20:41.345539676 +0000 UTC m=+1521.856150691" observedRunningTime="2025-12-10 19:20:44.063707509 +0000 UTC m=+1524.574318524" watchObservedRunningTime="2025-12-10 19:20:44.065791644 +0000 UTC m=+1524.576402649" Dec 10 19:20:44 crc kubenswrapper[4828]: I1210 19:20:44.119900 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.119872628 podStartE2EDuration="5.119872628s" podCreationTimestamp="2025-12-10 19:20:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:44.085693738 +0000 UTC m=+1524.596304753" watchObservedRunningTime="2025-12-10 19:20:44.119872628 +0000 UTC m=+1524.630483633" Dec 10 19:20:44 crc kubenswrapper[4828]: I1210 19:20:44.545653 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 10 19:20:45 crc kubenswrapper[4828]: I1210 19:20:45.060451 4828 generic.go:334] "Generic (PLEG): container finished" podID="8cb3c5d9-61fb-415a-a644-75803d3f7c8f" containerID="00139c641c2eb6e20a80832cd0c631ec70582d91abe9adaad0315426b3e6248a" exitCode=143 Dec 10 19:20:45 crc kubenswrapper[4828]: I1210 19:20:45.060550 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8cb3c5d9-61fb-415a-a644-75803d3f7c8f","Type":"ContainerDied","Data":"00139c641c2eb6e20a80832cd0c631ec70582d91abe9adaad0315426b3e6248a"} Dec 10 19:20:45 crc kubenswrapper[4828]: I1210 19:20:45.829402 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6f465c7bf6-5f8bw" Dec 10 19:20:46 crc kubenswrapper[4828]: I1210 19:20:46.079690 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01a17451-e170-42ad-9082-d91eb11e237a","Type":"ContainerStarted","Data":"3d7b46df6fba141148393519f246fb75354b03e9874c1e63610825d8b173d618"} Dec 10 19:20:46 crc kubenswrapper[4828]: I1210 19:20:46.542376 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-56c96d9477-vps8l" Dec 10 19:20:46 crc kubenswrapper[4828]: I1210 19:20:46.627285 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-66756f49f6-r2lds"] Dec 10 19:20:46 crc kubenswrapper[4828]: I1210 19:20:46.627508 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-66756f49f6-r2lds" podUID="ddf32ed2-2731-497c-a021-6a00f495265f" containerName="neutron-api" containerID="cri-o://8a66efac7a7e4f9bc7e860f04d479420bce2ca5ba3a1e8f3e01a2fa150047e11" gracePeriod=30 Dec 10 19:20:46 crc kubenswrapper[4828]: I1210 19:20:46.627603 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-66756f49f6-r2lds" podUID="ddf32ed2-2731-497c-a021-6a00f495265f" containerName="neutron-httpd" containerID="cri-o://300ec2101e4f20386ba3f3b3c21d97e458deb77fd9598c06b4391b95209e458c" gracePeriod=30 Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.144439 4828 generic.go:334] "Generic (PLEG): container finished" podID="ddf32ed2-2731-497c-a021-6a00f495265f" containerID="300ec2101e4f20386ba3f3b3c21d97e458deb77fd9598c06b4391b95209e458c" exitCode=0 Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.145106 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66756f49f6-r2lds" event={"ID":"ddf32ed2-2731-497c-a021-6a00f495265f","Type":"ContainerDied","Data":"300ec2101e4f20386ba3f3b3c21d97e458deb77fd9598c06b4391b95209e458c"} Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.713645 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.715915 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.718361 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.718969 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.720257 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-sklq7" Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.727204 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.797932 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28cb9eb3-e89d-460c-9110-32e241d2a44d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"28cb9eb3-e89d-460c-9110-32e241d2a44d\") " pod="openstack/openstackclient" Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.798033 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/28cb9eb3-e89d-460c-9110-32e241d2a44d-openstack-config\") pod \"openstackclient\" (UID: \"28cb9eb3-e89d-460c-9110-32e241d2a44d\") " pod="openstack/openstackclient" Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.798096 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vntcq\" (UniqueName: \"kubernetes.io/projected/28cb9eb3-e89d-460c-9110-32e241d2a44d-kube-api-access-vntcq\") pod \"openstackclient\" (UID: \"28cb9eb3-e89d-460c-9110-32e241d2a44d\") " pod="openstack/openstackclient" Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.798119 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/28cb9eb3-e89d-460c-9110-32e241d2a44d-openstack-config-secret\") pod \"openstackclient\" (UID: \"28cb9eb3-e89d-460c-9110-32e241d2a44d\") " pod="openstack/openstackclient" Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.900575 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/28cb9eb3-e89d-460c-9110-32e241d2a44d-openstack-config\") pod \"openstackclient\" (UID: \"28cb9eb3-e89d-460c-9110-32e241d2a44d\") " pod="openstack/openstackclient" Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.900884 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vntcq\" (UniqueName: \"kubernetes.io/projected/28cb9eb3-e89d-460c-9110-32e241d2a44d-kube-api-access-vntcq\") pod \"openstackclient\" (UID: \"28cb9eb3-e89d-460c-9110-32e241d2a44d\") " pod="openstack/openstackclient" Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.900974 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/28cb9eb3-e89d-460c-9110-32e241d2a44d-openstack-config-secret\") pod \"openstackclient\" (UID: \"28cb9eb3-e89d-460c-9110-32e241d2a44d\") " pod="openstack/openstackclient" Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.901148 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28cb9eb3-e89d-460c-9110-32e241d2a44d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"28cb9eb3-e89d-460c-9110-32e241d2a44d\") " pod="openstack/openstackclient" Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.902485 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/28cb9eb3-e89d-460c-9110-32e241d2a44d-openstack-config\") pod \"openstackclient\" (UID: \"28cb9eb3-e89d-460c-9110-32e241d2a44d\") " pod="openstack/openstackclient" Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.908363 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/28cb9eb3-e89d-460c-9110-32e241d2a44d-openstack-config-secret\") pod \"openstackclient\" (UID: \"28cb9eb3-e89d-460c-9110-32e241d2a44d\") " pod="openstack/openstackclient" Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.908957 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28cb9eb3-e89d-460c-9110-32e241d2a44d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"28cb9eb3-e89d-460c-9110-32e241d2a44d\") " pod="openstack/openstackclient" Dec 10 19:20:47 crc kubenswrapper[4828]: I1210 19:20:47.919355 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vntcq\" (UniqueName: \"kubernetes.io/projected/28cb9eb3-e89d-460c-9110-32e241d2a44d-kube-api-access-vntcq\") pod \"openstackclient\" (UID: \"28cb9eb3-e89d-460c-9110-32e241d2a44d\") " pod="openstack/openstackclient" Dec 10 19:20:48 crc kubenswrapper[4828]: I1210 19:20:48.049512 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 10 19:20:48 crc kubenswrapper[4828]: I1210 19:20:48.166428 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01a17451-e170-42ad-9082-d91eb11e237a","Type":"ContainerStarted","Data":"4b1e30bb14a3c5882d416795ef6afe70e2bae7117078dd92b6b8f1fd151fc78a"} Dec 10 19:20:48 crc kubenswrapper[4828]: I1210 19:20:48.740113 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 10 19:20:48 crc kubenswrapper[4828]: W1210 19:20:48.753018 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28cb9eb3_e89d_460c_9110_32e241d2a44d.slice/crio-c2a7f8d88f0df3198732850f3356ac050336ce5b1f2f43ac4137a9acc7f0e47b WatchSource:0}: Error finding container c2a7f8d88f0df3198732850f3356ac050336ce5b1f2f43ac4137a9acc7f0e47b: Status 404 returned error can't find the container with id c2a7f8d88f0df3198732850f3356ac050336ce5b1f2f43ac4137a9acc7f0e47b Dec 10 19:20:49 crc kubenswrapper[4828]: I1210 19:20:49.178630 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"28cb9eb3-e89d-460c-9110-32e241d2a44d","Type":"ContainerStarted","Data":"c2a7f8d88f0df3198732850f3356ac050336ce5b1f2f43ac4137a9acc7f0e47b"} Dec 10 19:20:49 crc kubenswrapper[4828]: I1210 19:20:49.181217 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01a17451-e170-42ad-9082-d91eb11e237a","Type":"ContainerStarted","Data":"f6176ab9eb78a7828354e6fc0062a0829cccd9dd942af8ae835ba5b60ee40f22"} Dec 10 19:20:49 crc kubenswrapper[4828]: I1210 19:20:49.414071 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:20:49 crc kubenswrapper[4828]: I1210 19:20:49.485167 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xmnbf" podUID="5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" containerName="registry-server" probeResult="failure" output=< Dec 10 19:20:49 crc kubenswrapper[4828]: timeout: failed to connect service ":50051" within 1s Dec 10 19:20:49 crc kubenswrapper[4828]: > Dec 10 19:20:49 crc kubenswrapper[4828]: I1210 19:20:49.501950 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-pxfhx"] Dec 10 19:20:49 crc kubenswrapper[4828]: I1210 19:20:49.502203 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" podUID="ef7ea83b-99e3-4460-b3db-63b0fae43070" containerName="dnsmasq-dns" containerID="cri-o://cb9f5250a01313095b6980a1c075687600d887952e18da51cd0f2dba40d40fb0" gracePeriod=10 Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.207494 4828 generic.go:334] "Generic (PLEG): container finished" podID="ef7ea83b-99e3-4460-b3db-63b0fae43070" containerID="cb9f5250a01313095b6980a1c075687600d887952e18da51cd0f2dba40d40fb0" exitCode=0 Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.207534 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" event={"ID":"ef7ea83b-99e3-4460-b3db-63b0fae43070","Type":"ContainerDied","Data":"cb9f5250a01313095b6980a1c075687600d887952e18da51cd0f2dba40d40fb0"} Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.306095 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.428349 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.443261 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.473389 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-ovsdbserver-nb\") pod \"ef7ea83b-99e3-4460-b3db-63b0fae43070\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.473463 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-dns-svc\") pod \"ef7ea83b-99e3-4460-b3db-63b0fae43070\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.473522 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-ovsdbserver-sb\") pod \"ef7ea83b-99e3-4460-b3db-63b0fae43070\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.473551 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-config\") pod \"ef7ea83b-99e3-4460-b3db-63b0fae43070\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.473636 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-dns-swift-storage-0\") pod \"ef7ea83b-99e3-4460-b3db-63b0fae43070\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.473672 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4jvx\" (UniqueName: \"kubernetes.io/projected/ef7ea83b-99e3-4460-b3db-63b0fae43070-kube-api-access-g4jvx\") pod \"ef7ea83b-99e3-4460-b3db-63b0fae43070\" (UID: \"ef7ea83b-99e3-4460-b3db-63b0fae43070\") " Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.482637 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef7ea83b-99e3-4460-b3db-63b0fae43070-kube-api-access-g4jvx" (OuterVolumeSpecName: "kube-api-access-g4jvx") pod "ef7ea83b-99e3-4460-b3db-63b0fae43070" (UID: "ef7ea83b-99e3-4460-b3db-63b0fae43070"). InnerVolumeSpecName "kube-api-access-g4jvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.577235 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4jvx\" (UniqueName: \"kubernetes.io/projected/ef7ea83b-99e3-4460-b3db-63b0fae43070-kube-api-access-g4jvx\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.618407 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ef7ea83b-99e3-4460-b3db-63b0fae43070" (UID: "ef7ea83b-99e3-4460-b3db-63b0fae43070"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.625928 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ef7ea83b-99e3-4460-b3db-63b0fae43070" (UID: "ef7ea83b-99e3-4460-b3db-63b0fae43070"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.661004 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-config" (OuterVolumeSpecName: "config") pod "ef7ea83b-99e3-4460-b3db-63b0fae43070" (UID: "ef7ea83b-99e3-4460-b3db-63b0fae43070"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.661107 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ef7ea83b-99e3-4460-b3db-63b0fae43070" (UID: "ef7ea83b-99e3-4460-b3db-63b0fae43070"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.667266 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ef7ea83b-99e3-4460-b3db-63b0fae43070" (UID: "ef7ea83b-99e3-4460-b3db-63b0fae43070"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.682044 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.682078 4828 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.682087 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.682096 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:50 crc kubenswrapper[4828]: I1210 19:20:50.682104 4828 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef7ea83b-99e3-4460-b3db-63b0fae43070-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.227429 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" event={"ID":"ef7ea83b-99e3-4460-b3db-63b0fae43070","Type":"ContainerDied","Data":"87da3edc14147c6aa655cd36b8561b454ac562daef1a885e756fe9087bdb62e6"} Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.227496 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-pxfhx" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.227920 4828 scope.go:117] "RemoveContainer" containerID="cb9f5250a01313095b6980a1c075687600d887952e18da51cd0f2dba40d40fb0" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.230031 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.230074 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.230120 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.231090 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.231151 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" gracePeriod=600 Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.235520 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01a17451-e170-42ad-9082-d91eb11e237a","Type":"ContainerStarted","Data":"978bc8cc5cfbb1ea03b6dd0f9e5ed2ce4a7dd7ef8fac884740f815c131552c14"} Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.235806 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.242316 4828 generic.go:334] "Generic (PLEG): container finished" podID="ddf32ed2-2731-497c-a021-6a00f495265f" containerID="8a66efac7a7e4f9bc7e860f04d479420bce2ca5ba3a1e8f3e01a2fa150047e11" exitCode=0 Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.242478 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66756f49f6-r2lds" event={"ID":"ddf32ed2-2731-497c-a021-6a00f495265f","Type":"ContainerDied","Data":"8a66efac7a7e4f9bc7e860f04d479420bce2ca5ba3a1e8f3e01a2fa150047e11"} Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.242575 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="c5297488-c2cd-4c22-906b-2df48517ff45" containerName="cinder-scheduler" containerID="cri-o://99f10991c34d4247de761b42a0409fa7fb664ea2b7f9bde6181574e2257aa8b7" gracePeriod=30 Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.242719 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="c5297488-c2cd-4c22-906b-2df48517ff45" containerName="probe" containerID="cri-o://173bd8216ca59a312cafa05f0f282643d70e91472d040004aae16edd4c800577" gracePeriod=30 Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.284238 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.645152622 podStartE2EDuration="9.284213072s" podCreationTimestamp="2025-12-10 19:20:42 +0000 UTC" firstStartedPulling="2025-12-10 19:20:43.281937561 +0000 UTC m=+1523.792548566" lastFinishedPulling="2025-12-10 19:20:49.920998011 +0000 UTC m=+1530.431609016" observedRunningTime="2025-12-10 19:20:51.27352602 +0000 UTC m=+1531.784137025" watchObservedRunningTime="2025-12-10 19:20:51.284213072 +0000 UTC m=+1531.794824077" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.319525 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-pxfhx"] Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.333100 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-pxfhx"] Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.336834 4828 scope.go:117] "RemoveContainer" containerID="e15eddcbf82a51fcee4bed867488548b482aa968109b31d4dad7526bb32b85fa" Dec 10 19:20:51 crc kubenswrapper[4828]: E1210 19:20:51.358556 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.536234 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.602936 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-combined-ca-bundle\") pod \"ddf32ed2-2731-497c-a021-6a00f495265f\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.603719 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-ovndb-tls-certs\") pod \"ddf32ed2-2731-497c-a021-6a00f495265f\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.603886 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-httpd-config\") pod \"ddf32ed2-2731-497c-a021-6a00f495265f\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.603987 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvc92\" (UniqueName: \"kubernetes.io/projected/ddf32ed2-2731-497c-a021-6a00f495265f-kube-api-access-xvc92\") pod \"ddf32ed2-2731-497c-a021-6a00f495265f\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.604189 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-config\") pod \"ddf32ed2-2731-497c-a021-6a00f495265f\" (UID: \"ddf32ed2-2731-497c-a021-6a00f495265f\") " Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.613976 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "ddf32ed2-2731-497c-a021-6a00f495265f" (UID: "ddf32ed2-2731-497c-a021-6a00f495265f"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.648377 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddf32ed2-2731-497c-a021-6a00f495265f-kube-api-access-xvc92" (OuterVolumeSpecName: "kube-api-access-xvc92") pod "ddf32ed2-2731-497c-a021-6a00f495265f" (UID: "ddf32ed2-2731-497c-a021-6a00f495265f"). InnerVolumeSpecName "kube-api-access-xvc92". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.710228 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-config" (OuterVolumeSpecName: "config") pod "ddf32ed2-2731-497c-a021-6a00f495265f" (UID: "ddf32ed2-2731-497c-a021-6a00f495265f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.710441 4828 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.710971 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvc92\" (UniqueName: \"kubernetes.io/projected/ddf32ed2-2731-497c-a021-6a00f495265f-kube-api-access-xvc92\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.722793 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ddf32ed2-2731-497c-a021-6a00f495265f" (UID: "ddf32ed2-2731-497c-a021-6a00f495265f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.755485 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "ddf32ed2-2731-497c-a021-6a00f495265f" (UID: "ddf32ed2-2731-497c-a021-6a00f495265f"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.814343 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef7ea83b-99e3-4460-b3db-63b0fae43070" path="/var/lib/kubelet/pods/ef7ea83b-99e3-4460-b3db-63b0fae43070/volumes" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.815325 4828 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.815351 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:51 crc kubenswrapper[4828]: I1210 19:20:51.815362 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddf32ed2-2731-497c-a021-6a00f495265f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:52 crc kubenswrapper[4828]: I1210 19:20:52.264245 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66756f49f6-r2lds" event={"ID":"ddf32ed2-2731-497c-a021-6a00f495265f","Type":"ContainerDied","Data":"149e691def5b33b1a1a967d48f2a38ebd954ff995beab863116acedb67250efc"} Dec 10 19:20:52 crc kubenswrapper[4828]: I1210 19:20:52.264294 4828 scope.go:117] "RemoveContainer" containerID="300ec2101e4f20386ba3f3b3c21d97e458deb77fd9598c06b4391b95209e458c" Dec 10 19:20:52 crc kubenswrapper[4828]: I1210 19:20:52.264426 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66756f49f6-r2lds" Dec 10 19:20:52 crc kubenswrapper[4828]: I1210 19:20:52.269074 4828 generic.go:334] "Generic (PLEG): container finished" podID="c5297488-c2cd-4c22-906b-2df48517ff45" containerID="173bd8216ca59a312cafa05f0f282643d70e91472d040004aae16edd4c800577" exitCode=0 Dec 10 19:20:52 crc kubenswrapper[4828]: I1210 19:20:52.269159 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c5297488-c2cd-4c22-906b-2df48517ff45","Type":"ContainerDied","Data":"173bd8216ca59a312cafa05f0f282643d70e91472d040004aae16edd4c800577"} Dec 10 19:20:52 crc kubenswrapper[4828]: I1210 19:20:52.278603 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" exitCode=0 Dec 10 19:20:52 crc kubenswrapper[4828]: I1210 19:20:52.278657 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca"} Dec 10 19:20:52 crc kubenswrapper[4828]: I1210 19:20:52.279856 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:20:52 crc kubenswrapper[4828]: E1210 19:20:52.280152 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:20:52 crc kubenswrapper[4828]: I1210 19:20:52.295155 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-66756f49f6-r2lds"] Dec 10 19:20:52 crc kubenswrapper[4828]: I1210 19:20:52.310078 4828 scope.go:117] "RemoveContainer" containerID="8a66efac7a7e4f9bc7e860f04d479420bce2ca5ba3a1e8f3e01a2fa150047e11" Dec 10 19:20:52 crc kubenswrapper[4828]: I1210 19:20:52.319303 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-66756f49f6-r2lds"] Dec 10 19:20:52 crc kubenswrapper[4828]: I1210 19:20:52.338974 4828 scope.go:117] "RemoveContainer" containerID="04946921ee0aa69a9d56354bf922c69f57770e0a2b673248523db9ef4278fd20" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.058602 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.068224 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.141350 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-config-data\") pod \"c5297488-c2cd-4c22-906b-2df48517ff45\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.141490 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-scripts\") pod \"c5297488-c2cd-4c22-906b-2df48517ff45\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.141570 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5297488-c2cd-4c22-906b-2df48517ff45-etc-machine-id\") pod \"c5297488-c2cd-4c22-906b-2df48517ff45\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.141651 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kj652\" (UniqueName: \"kubernetes.io/projected/c5297488-c2cd-4c22-906b-2df48517ff45-kube-api-access-kj652\") pod \"c5297488-c2cd-4c22-906b-2df48517ff45\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.141675 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-combined-ca-bundle\") pod \"c5297488-c2cd-4c22-906b-2df48517ff45\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.141768 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-config-data-custom\") pod \"c5297488-c2cd-4c22-906b-2df48517ff45\" (UID: \"c5297488-c2cd-4c22-906b-2df48517ff45\") " Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.141893 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5297488-c2cd-4c22-906b-2df48517ff45-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c5297488-c2cd-4c22-906b-2df48517ff45" (UID: "c5297488-c2cd-4c22-906b-2df48517ff45"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.142180 4828 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c5297488-c2cd-4c22-906b-2df48517ff45-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.152040 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-scripts" (OuterVolumeSpecName: "scripts") pod "c5297488-c2cd-4c22-906b-2df48517ff45" (UID: "c5297488-c2cd-4c22-906b-2df48517ff45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.152384 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5297488-c2cd-4c22-906b-2df48517ff45-kube-api-access-kj652" (OuterVolumeSpecName: "kube-api-access-kj652") pod "c5297488-c2cd-4c22-906b-2df48517ff45" (UID: "c5297488-c2cd-4c22-906b-2df48517ff45"). InnerVolumeSpecName "kube-api-access-kj652". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.152850 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c5297488-c2cd-4c22-906b-2df48517ff45" (UID: "c5297488-c2cd-4c22-906b-2df48517ff45"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.243910 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kj652\" (UniqueName: \"kubernetes.io/projected/c5297488-c2cd-4c22-906b-2df48517ff45-kube-api-access-kj652\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.243937 4828 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.243946 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.275691 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5297488-c2cd-4c22-906b-2df48517ff45" (UID: "c5297488-c2cd-4c22-906b-2df48517ff45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.319510 4828 generic.go:334] "Generic (PLEG): container finished" podID="c5297488-c2cd-4c22-906b-2df48517ff45" containerID="99f10991c34d4247de761b42a0409fa7fb664ea2b7f9bde6181574e2257aa8b7" exitCode=0 Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.319566 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c5297488-c2cd-4c22-906b-2df48517ff45","Type":"ContainerDied","Data":"99f10991c34d4247de761b42a0409fa7fb664ea2b7f9bde6181574e2257aa8b7"} Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.319630 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c5297488-c2cd-4c22-906b-2df48517ff45","Type":"ContainerDied","Data":"cb51f1d75749de9a2640714f4b5c7523910d452007200eb5fd3d66f70bb4f070"} Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.319652 4828 scope.go:117] "RemoveContainer" containerID="173bd8216ca59a312cafa05f0f282643d70e91472d040004aae16edd4c800577" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.319789 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.349498 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-config-data" (OuterVolumeSpecName: "config-data") pod "c5297488-c2cd-4c22-906b-2df48517ff45" (UID: "c5297488-c2cd-4c22-906b-2df48517ff45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.363500 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.363741 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5297488-c2cd-4c22-906b-2df48517ff45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.474968 4828 scope.go:117] "RemoveContainer" containerID="99f10991c34d4247de761b42a0409fa7fb664ea2b7f9bde6181574e2257aa8b7" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.532937 4828 scope.go:117] "RemoveContainer" containerID="173bd8216ca59a312cafa05f0f282643d70e91472d040004aae16edd4c800577" Dec 10 19:20:53 crc kubenswrapper[4828]: E1210 19:20:53.549650 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"173bd8216ca59a312cafa05f0f282643d70e91472d040004aae16edd4c800577\": container with ID starting with 173bd8216ca59a312cafa05f0f282643d70e91472d040004aae16edd4c800577 not found: ID does not exist" containerID="173bd8216ca59a312cafa05f0f282643d70e91472d040004aae16edd4c800577" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.549701 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"173bd8216ca59a312cafa05f0f282643d70e91472d040004aae16edd4c800577"} err="failed to get container status \"173bd8216ca59a312cafa05f0f282643d70e91472d040004aae16edd4c800577\": rpc error: code = NotFound desc = could not find container \"173bd8216ca59a312cafa05f0f282643d70e91472d040004aae16edd4c800577\": container with ID starting with 173bd8216ca59a312cafa05f0f282643d70e91472d040004aae16edd4c800577 not found: ID does not exist" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.549732 4828 scope.go:117] "RemoveContainer" containerID="99f10991c34d4247de761b42a0409fa7fb664ea2b7f9bde6181574e2257aa8b7" Dec 10 19:20:53 crc kubenswrapper[4828]: E1210 19:20:53.566831 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99f10991c34d4247de761b42a0409fa7fb664ea2b7f9bde6181574e2257aa8b7\": container with ID starting with 99f10991c34d4247de761b42a0409fa7fb664ea2b7f9bde6181574e2257aa8b7 not found: ID does not exist" containerID="99f10991c34d4247de761b42a0409fa7fb664ea2b7f9bde6181574e2257aa8b7" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.566871 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99f10991c34d4247de761b42a0409fa7fb664ea2b7f9bde6181574e2257aa8b7"} err="failed to get container status \"99f10991c34d4247de761b42a0409fa7fb664ea2b7f9bde6181574e2257aa8b7\": rpc error: code = NotFound desc = could not find container \"99f10991c34d4247de761b42a0409fa7fb664ea2b7f9bde6181574e2257aa8b7\": container with ID starting with 99f10991c34d4247de761b42a0409fa7fb664ea2b7f9bde6181574e2257aa8b7 not found: ID does not exist" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.661862 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.676388 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.689773 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:20:53 crc kubenswrapper[4828]: E1210 19:20:53.690299 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef7ea83b-99e3-4460-b3db-63b0fae43070" containerName="init" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.690319 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef7ea83b-99e3-4460-b3db-63b0fae43070" containerName="init" Dec 10 19:20:53 crc kubenswrapper[4828]: E1210 19:20:53.690332 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5297488-c2cd-4c22-906b-2df48517ff45" containerName="cinder-scheduler" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.690339 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5297488-c2cd-4c22-906b-2df48517ff45" containerName="cinder-scheduler" Dec 10 19:20:53 crc kubenswrapper[4828]: E1210 19:20:53.690364 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef7ea83b-99e3-4460-b3db-63b0fae43070" containerName="dnsmasq-dns" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.690370 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef7ea83b-99e3-4460-b3db-63b0fae43070" containerName="dnsmasq-dns" Dec 10 19:20:53 crc kubenswrapper[4828]: E1210 19:20:53.690391 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5297488-c2cd-4c22-906b-2df48517ff45" containerName="probe" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.690396 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5297488-c2cd-4c22-906b-2df48517ff45" containerName="probe" Dec 10 19:20:53 crc kubenswrapper[4828]: E1210 19:20:53.690406 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddf32ed2-2731-497c-a021-6a00f495265f" containerName="neutron-api" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.690413 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddf32ed2-2731-497c-a021-6a00f495265f" containerName="neutron-api" Dec 10 19:20:53 crc kubenswrapper[4828]: E1210 19:20:53.690439 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddf32ed2-2731-497c-a021-6a00f495265f" containerName="neutron-httpd" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.690445 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddf32ed2-2731-497c-a021-6a00f495265f" containerName="neutron-httpd" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.690644 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5297488-c2cd-4c22-906b-2df48517ff45" containerName="probe" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.690657 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef7ea83b-99e3-4460-b3db-63b0fae43070" containerName="dnsmasq-dns" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.690675 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddf32ed2-2731-497c-a021-6a00f495265f" containerName="neutron-api" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.690688 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddf32ed2-2731-497c-a021-6a00f495265f" containerName="neutron-httpd" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.690701 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5297488-c2cd-4c22-906b-2df48517ff45" containerName="cinder-scheduler" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.691939 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.694317 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.734297 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.774919 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f65c47b-68bb-4c64-b0e8-94507fb828e5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.774965 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f65c47b-68bb-4c64-b0e8-94507fb828e5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.775001 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf7pb\" (UniqueName: \"kubernetes.io/projected/8f65c47b-68bb-4c64-b0e8-94507fb828e5-kube-api-access-xf7pb\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.775021 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f65c47b-68bb-4c64-b0e8-94507fb828e5-config-data\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.775144 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f65c47b-68bb-4c64-b0e8-94507fb828e5-scripts\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.775180 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f65c47b-68bb-4c64-b0e8-94507fb828e5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.810039 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5297488-c2cd-4c22-906b-2df48517ff45" path="/var/lib/kubelet/pods/c5297488-c2cd-4c22-906b-2df48517ff45/volumes" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.810615 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddf32ed2-2731-497c-a021-6a00f495265f" path="/var/lib/kubelet/pods/ddf32ed2-2731-497c-a021-6a00f495265f/volumes" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.877485 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f65c47b-68bb-4c64-b0e8-94507fb828e5-scripts\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.877594 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f65c47b-68bb-4c64-b0e8-94507fb828e5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.877744 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f65c47b-68bb-4c64-b0e8-94507fb828e5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.877774 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f65c47b-68bb-4c64-b0e8-94507fb828e5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.877854 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf7pb\" (UniqueName: \"kubernetes.io/projected/8f65c47b-68bb-4c64-b0e8-94507fb828e5-kube-api-access-xf7pb\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.877909 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f65c47b-68bb-4c64-b0e8-94507fb828e5-config-data\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.877903 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f65c47b-68bb-4c64-b0e8-94507fb828e5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.884601 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f65c47b-68bb-4c64-b0e8-94507fb828e5-config-data\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.885783 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f65c47b-68bb-4c64-b0e8-94507fb828e5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.885912 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f65c47b-68bb-4c64-b0e8-94507fb828e5-scripts\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.887304 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f65c47b-68bb-4c64-b0e8-94507fb828e5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:53 crc kubenswrapper[4828]: I1210 19:20:53.899294 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf7pb\" (UniqueName: \"kubernetes.io/projected/8f65c47b-68bb-4c64-b0e8-94507fb828e5-kube-api-access-xf7pb\") pod \"cinder-scheduler-0\" (UID: \"8f65c47b-68bb-4c64-b0e8-94507fb828e5\") " pod="openstack/cinder-scheduler-0" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.029254 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.454108 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7b757664c-hkcjf"] Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.466190 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7b757664c-hkcjf"] Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.466373 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.477779 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.479206 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.479561 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.534236 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.605166 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-etc-swift\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.605293 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-run-httpd\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.605427 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-internal-tls-certs\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.605575 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmtdn\" (UniqueName: \"kubernetes.io/projected/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-kube-api-access-lmtdn\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.605714 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-log-httpd\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.605824 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-combined-ca-bundle\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.605841 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-public-tls-certs\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.605865 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-config-data\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.707644 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-internal-tls-certs\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.707708 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmtdn\" (UniqueName: \"kubernetes.io/projected/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-kube-api-access-lmtdn\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.707731 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-log-httpd\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.707773 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-combined-ca-bundle\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.707788 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-public-tls-certs\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.707828 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-config-data\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.707934 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-etc-swift\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.707970 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-run-httpd\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.708484 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-run-httpd\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.710290 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-log-httpd\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.714387 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-config-data\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.715924 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-combined-ca-bundle\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.716273 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-internal-tls-certs\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.717669 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-public-tls-certs\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.721002 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-etc-swift\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.729866 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmtdn\" (UniqueName: \"kubernetes.io/projected/c6749e4d-e98e-441c-a1e8-2a1ada77eb84-kube-api-access-lmtdn\") pod \"swift-proxy-7b757664c-hkcjf\" (UID: \"c6749e4d-e98e-441c-a1e8-2a1ada77eb84\") " pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.801281 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.915023 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-787d69597b-vd29s"] Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.917307 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.920205 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.920506 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-jrdmz" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.920664 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 10 19:20:54 crc kubenswrapper[4828]: I1210 19:20:54.961989 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-787d69597b-vd29s"] Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.024213 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-config-data\") pod \"heat-engine-787d69597b-vd29s\" (UID: \"abb273e9-4472-4b35-bd92-5c00f7613bd7\") " pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.024320 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-combined-ca-bundle\") pod \"heat-engine-787d69597b-vd29s\" (UID: \"abb273e9-4472-4b35-bd92-5c00f7613bd7\") " pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.024381 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-config-data-custom\") pod \"heat-engine-787d69597b-vd29s\" (UID: \"abb273e9-4472-4b35-bd92-5c00f7613bd7\") " pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.024513 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7pk8\" (UniqueName: \"kubernetes.io/projected/abb273e9-4472-4b35-bd92-5c00f7613bd7-kube-api-access-m7pk8\") pod \"heat-engine-787d69597b-vd29s\" (UID: \"abb273e9-4472-4b35-bd92-5c00f7613bd7\") " pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.046301 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-6f7b6d44ff-rhjwx"] Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.047857 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.060844 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.098144 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6f7b6d44ff-rhjwx"] Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.135039 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-jlhgg"] Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.136950 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.154051 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-combined-ca-bundle\") pod \"heat-engine-787d69597b-vd29s\" (UID: \"abb273e9-4472-4b35-bd92-5c00f7613bd7\") " pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.154187 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-config-data-custom\") pod \"heat-engine-787d69597b-vd29s\" (UID: \"abb273e9-4472-4b35-bd92-5c00f7613bd7\") " pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.154427 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7pk8\" (UniqueName: \"kubernetes.io/projected/abb273e9-4472-4b35-bd92-5c00f7613bd7-kube-api-access-m7pk8\") pod \"heat-engine-787d69597b-vd29s\" (UID: \"abb273e9-4472-4b35-bd92-5c00f7613bd7\") " pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.154546 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-combined-ca-bundle\") pod \"heat-cfnapi-6f7b6d44ff-rhjwx\" (UID: \"51c89612-b937-4376-88a8-6623b3ec8d05\") " pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.154642 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-config-data\") pod \"heat-cfnapi-6f7b6d44ff-rhjwx\" (UID: \"51c89612-b937-4376-88a8-6623b3ec8d05\") " pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.154688 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jmgb\" (UniqueName: \"kubernetes.io/projected/51c89612-b937-4376-88a8-6623b3ec8d05-kube-api-access-5jmgb\") pod \"heat-cfnapi-6f7b6d44ff-rhjwx\" (UID: \"51c89612-b937-4376-88a8-6623b3ec8d05\") " pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.154712 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-config-data\") pod \"heat-engine-787d69597b-vd29s\" (UID: \"abb273e9-4472-4b35-bd92-5c00f7613bd7\") " pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.154826 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-config-data-custom\") pod \"heat-cfnapi-6f7b6d44ff-rhjwx\" (UID: \"51c89612-b937-4376-88a8-6623b3ec8d05\") " pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.158081 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-jlhgg"] Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.171128 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-config-data-custom\") pod \"heat-engine-787d69597b-vd29s\" (UID: \"abb273e9-4472-4b35-bd92-5c00f7613bd7\") " pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.182108 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-config-data\") pod \"heat-engine-787d69597b-vd29s\" (UID: \"abb273e9-4472-4b35-bd92-5c00f7613bd7\") " pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.191316 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-combined-ca-bundle\") pod \"heat-engine-787d69597b-vd29s\" (UID: \"abb273e9-4472-4b35-bd92-5c00f7613bd7\") " pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.230455 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7pk8\" (UniqueName: \"kubernetes.io/projected/abb273e9-4472-4b35-bd92-5c00f7613bd7-kube-api-access-m7pk8\") pod \"heat-engine-787d69597b-vd29s\" (UID: \"abb273e9-4472-4b35-bd92-5c00f7613bd7\") " pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.256280 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-77bd4457c-f5tqz"] Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.257993 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-77bd4457c-f5tqz" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.261046 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-combined-ca-bundle\") pod \"heat-cfnapi-6f7b6d44ff-rhjwx\" (UID: \"51c89612-b937-4376-88a8-6623b3ec8d05\") " pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.261116 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-config-data\") pod \"heat-cfnapi-6f7b6d44ff-rhjwx\" (UID: \"51c89612-b937-4376-88a8-6623b3ec8d05\") " pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.261146 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jmgb\" (UniqueName: \"kubernetes.io/projected/51c89612-b937-4376-88a8-6623b3ec8d05-kube-api-access-5jmgb\") pod \"heat-cfnapi-6f7b6d44ff-rhjwx\" (UID: \"51c89612-b937-4376-88a8-6623b3ec8d05\") " pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.261208 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-config-data-custom\") pod \"heat-cfnapi-6f7b6d44ff-rhjwx\" (UID: \"51c89612-b937-4376-88a8-6623b3ec8d05\") " pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.268085 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.281212 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-combined-ca-bundle\") pod \"heat-cfnapi-6f7b6d44ff-rhjwx\" (UID: \"51c89612-b937-4376-88a8-6623b3ec8d05\") " pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.282352 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-config-data\") pod \"heat-cfnapi-6f7b6d44ff-rhjwx\" (UID: \"51c89612-b937-4376-88a8-6623b3ec8d05\") " pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.299827 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-config-data-custom\") pod \"heat-cfnapi-6f7b6d44ff-rhjwx\" (UID: \"51c89612-b937-4376-88a8-6623b3ec8d05\") " pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.307756 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jmgb\" (UniqueName: \"kubernetes.io/projected/51c89612-b937-4376-88a8-6623b3ec8d05-kube-api-access-5jmgb\") pod \"heat-cfnapi-6f7b6d44ff-rhjwx\" (UID: \"51c89612-b937-4376-88a8-6623b3ec8d05\") " pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.336517 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.340028 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-77bd4457c-f5tqz"] Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.423179 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.423281 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-config-data\") pod \"heat-api-77bd4457c-f5tqz\" (UID: \"6e68a634-d019-464c-bcf7-c1669cf88fcf\") " pod="openstack/heat-api-77bd4457c-f5tqz" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.434440 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.438057 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-combined-ca-bundle\") pod \"heat-api-77bd4457c-f5tqz\" (UID: \"6e68a634-d019-464c-bcf7-c1669cf88fcf\") " pod="openstack/heat-api-77bd4457c-f5tqz" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.438224 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgb7h\" (UniqueName: \"kubernetes.io/projected/41b1aa0c-723c-4f3f-9e52-26328c8ca954-kube-api-access-hgb7h\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.438249 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-config-data-custom\") pod \"heat-api-77bd4457c-f5tqz\" (UID: \"6e68a634-d019-464c-bcf7-c1669cf88fcf\") " pod="openstack/heat-api-77bd4457c-f5tqz" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.438380 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-config\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.438407 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.438447 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.438579 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.438666 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgr5d\" (UniqueName: \"kubernetes.io/projected/6e68a634-d019-464c-bcf7-c1669cf88fcf-kube-api-access-fgr5d\") pod \"heat-api-77bd4457c-f5tqz\" (UID: \"6e68a634-d019-464c-bcf7-c1669cf88fcf\") " pod="openstack/heat-api-77bd4457c-f5tqz" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.498370 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8f65c47b-68bb-4c64-b0e8-94507fb828e5","Type":"ContainerStarted","Data":"9a5858649a666ad89cd6b94319240f23108108fe3846e447fb2b3f264fad0782"} Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.544486 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.544578 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-config-data\") pod \"heat-api-77bd4457c-f5tqz\" (UID: \"6e68a634-d019-464c-bcf7-c1669cf88fcf\") " pod="openstack/heat-api-77bd4457c-f5tqz" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.544648 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-combined-ca-bundle\") pod \"heat-api-77bd4457c-f5tqz\" (UID: \"6e68a634-d019-464c-bcf7-c1669cf88fcf\") " pod="openstack/heat-api-77bd4457c-f5tqz" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.544716 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgb7h\" (UniqueName: \"kubernetes.io/projected/41b1aa0c-723c-4f3f-9e52-26328c8ca954-kube-api-access-hgb7h\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.544740 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-config-data-custom\") pod \"heat-api-77bd4457c-f5tqz\" (UID: \"6e68a634-d019-464c-bcf7-c1669cf88fcf\") " pod="openstack/heat-api-77bd4457c-f5tqz" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.545026 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-config\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.545054 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.545088 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.545171 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.545224 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgr5d\" (UniqueName: \"kubernetes.io/projected/6e68a634-d019-464c-bcf7-c1669cf88fcf-kube-api-access-fgr5d\") pod \"heat-api-77bd4457c-f5tqz\" (UID: \"6e68a634-d019-464c-bcf7-c1669cf88fcf\") " pod="openstack/heat-api-77bd4457c-f5tqz" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.557370 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-config-data\") pod \"heat-api-77bd4457c-f5tqz\" (UID: \"6e68a634-d019-464c-bcf7-c1669cf88fcf\") " pod="openstack/heat-api-77bd4457c-f5tqz" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.562250 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.562638 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-config\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.567342 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.567405 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.568154 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-config-data-custom\") pod \"heat-api-77bd4457c-f5tqz\" (UID: \"6e68a634-d019-464c-bcf7-c1669cf88fcf\") " pod="openstack/heat-api-77bd4457c-f5tqz" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.571152 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-combined-ca-bundle\") pod \"heat-api-77bd4457c-f5tqz\" (UID: \"6e68a634-d019-464c-bcf7-c1669cf88fcf\") " pod="openstack/heat-api-77bd4457c-f5tqz" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.571575 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.591110 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgr5d\" (UniqueName: \"kubernetes.io/projected/6e68a634-d019-464c-bcf7-c1669cf88fcf-kube-api-access-fgr5d\") pod \"heat-api-77bd4457c-f5tqz\" (UID: \"6e68a634-d019-464c-bcf7-c1669cf88fcf\") " pod="openstack/heat-api-77bd4457c-f5tqz" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.659610 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgb7h\" (UniqueName: \"kubernetes.io/projected/41b1aa0c-723c-4f3f-9e52-26328c8ca954-kube-api-access-hgb7h\") pod \"dnsmasq-dns-688b9f5b49-jlhgg\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.714260 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.734295 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-77bd4457c-f5tqz" Dec 10 19:20:55 crc kubenswrapper[4828]: I1210 19:20:55.892853 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7b757664c-hkcjf"] Dec 10 19:20:56 crc kubenswrapper[4828]: I1210 19:20:56.143714 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-787d69597b-vd29s"] Dec 10 19:20:56 crc kubenswrapper[4828]: I1210 19:20:56.433594 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6f7b6d44ff-rhjwx"] Dec 10 19:20:56 crc kubenswrapper[4828]: I1210 19:20:56.557442 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b757664c-hkcjf" event={"ID":"c6749e4d-e98e-441c-a1e8-2a1ada77eb84","Type":"ContainerStarted","Data":"fe9a4ee67220ecab180131b0d01db901854ff9842324f064bdf8d63c0ae971f1"} Dec 10 19:20:56 crc kubenswrapper[4828]: I1210 19:20:56.565129 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-787d69597b-vd29s" event={"ID":"abb273e9-4472-4b35-bd92-5c00f7613bd7","Type":"ContainerStarted","Data":"c75a3dd941554e1df1af3c6b9fe656d1458ee4b582e8530c8c25da0e5c18476f"} Dec 10 19:20:57 crc kubenswrapper[4828]: I1210 19:20:57.162178 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-jlhgg"] Dec 10 19:20:57 crc kubenswrapper[4828]: W1210 19:20:57.348615 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41b1aa0c_723c_4f3f_9e52_26328c8ca954.slice/crio-2430290d21788f1db2b4c1e3acde5854f44e131999245d3af1e9b686411297d8 WatchSource:0}: Error finding container 2430290d21788f1db2b4c1e3acde5854f44e131999245d3af1e9b686411297d8: Status 404 returned error can't find the container with id 2430290d21788f1db2b4c1e3acde5854f44e131999245d3af1e9b686411297d8 Dec 10 19:20:57 crc kubenswrapper[4828]: I1210 19:20:57.390097 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-77bd4457c-f5tqz"] Dec 10 19:20:57 crc kubenswrapper[4828]: I1210 19:20:57.610276 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" event={"ID":"41b1aa0c-723c-4f3f-9e52-26328c8ca954","Type":"ContainerStarted","Data":"2430290d21788f1db2b4c1e3acde5854f44e131999245d3af1e9b686411297d8"} Dec 10 19:20:57 crc kubenswrapper[4828]: I1210 19:20:57.618983 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" event={"ID":"51c89612-b937-4376-88a8-6623b3ec8d05","Type":"ContainerStarted","Data":"2cead302fa78ebee00121c74029cb5f302dc7a3ce5f2c3981c0fd7adbc9b2634"} Dec 10 19:20:57 crc kubenswrapper[4828]: I1210 19:20:57.631000 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b757664c-hkcjf" event={"ID":"c6749e4d-e98e-441c-a1e8-2a1ada77eb84","Type":"ContainerStarted","Data":"477b1f691a76099478b2650b3e27c217c23bc82a23949fd557b9605f42a6b490"} Dec 10 19:20:57 crc kubenswrapper[4828]: I1210 19:20:57.660791 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-77bd4457c-f5tqz" event={"ID":"6e68a634-d019-464c-bcf7-c1669cf88fcf","Type":"ContainerStarted","Data":"77639af20d9aeb386bf2fa122b3d2ec2c301101e6cf5101fd48508a1e778363a"} Dec 10 19:20:57 crc kubenswrapper[4828]: I1210 19:20:57.671198 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-787d69597b-vd29s" event={"ID":"abb273e9-4472-4b35-bd92-5c00f7613bd7","Type":"ContainerStarted","Data":"fc35b017bc5e9c9d4621c6f6073a56b6b382e06001a2b8fcd5b8680ffd70a2d5"} Dec 10 19:20:57 crc kubenswrapper[4828]: I1210 19:20:57.672841 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:20:57 crc kubenswrapper[4828]: I1210 19:20:57.678852 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8f65c47b-68bb-4c64-b0e8-94507fb828e5","Type":"ContainerStarted","Data":"6b51048ed994de68a03222159e2f3bdc8fe8097fbede9c48219ed45fef7f4dd1"} Dec 10 19:20:57 crc kubenswrapper[4828]: I1210 19:20:57.689639 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-787d69597b-vd29s" podStartSLOduration=3.689621448 podStartE2EDuration="3.689621448s" podCreationTimestamp="2025-12-10 19:20:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:57.687769899 +0000 UTC m=+1538.198380904" watchObservedRunningTime="2025-12-10 19:20:57.689621448 +0000 UTC m=+1538.200232443" Dec 10 19:20:58 crc kubenswrapper[4828]: I1210 19:20:58.330697 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:20:58 crc kubenswrapper[4828]: I1210 19:20:58.334904 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01a17451-e170-42ad-9082-d91eb11e237a" containerName="ceilometer-central-agent" containerID="cri-o://3d7b46df6fba141148393519f246fb75354b03e9874c1e63610825d8b173d618" gracePeriod=30 Dec 10 19:20:58 crc kubenswrapper[4828]: I1210 19:20:58.337678 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01a17451-e170-42ad-9082-d91eb11e237a" containerName="proxy-httpd" containerID="cri-o://978bc8cc5cfbb1ea03b6dd0f9e5ed2ce4a7dd7ef8fac884740f815c131552c14" gracePeriod=30 Dec 10 19:20:58 crc kubenswrapper[4828]: I1210 19:20:58.337870 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01a17451-e170-42ad-9082-d91eb11e237a" containerName="sg-core" containerID="cri-o://f6176ab9eb78a7828354e6fc0062a0829cccd9dd942af8ae835ba5b60ee40f22" gracePeriod=30 Dec 10 19:20:58 crc kubenswrapper[4828]: I1210 19:20:58.337941 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="01a17451-e170-42ad-9082-d91eb11e237a" containerName="ceilometer-notification-agent" containerID="cri-o://4b1e30bb14a3c5882d416795ef6afe70e2bae7117078dd92b6b8f1fd151fc78a" gracePeriod=30 Dec 10 19:20:58 crc kubenswrapper[4828]: I1210 19:20:58.716923 4828 generic.go:334] "Generic (PLEG): container finished" podID="41b1aa0c-723c-4f3f-9e52-26328c8ca954" containerID="9be33178bb0a13451b3f987b140e3b14a6d716ce4cc174a6a394d6c6af0ef3bc" exitCode=0 Dec 10 19:20:58 crc kubenswrapper[4828]: I1210 19:20:58.718065 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" event={"ID":"41b1aa0c-723c-4f3f-9e52-26328c8ca954","Type":"ContainerDied","Data":"9be33178bb0a13451b3f987b140e3b14a6d716ce4cc174a6a394d6c6af0ef3bc"} Dec 10 19:20:58 crc kubenswrapper[4828]: I1210 19:20:58.762367 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7b757664c-hkcjf" event={"ID":"c6749e4d-e98e-441c-a1e8-2a1ada77eb84","Type":"ContainerStarted","Data":"d34f8a5db971870bd4873d8f47e451de3d4cfad4e8216abdfee235556b549fe0"} Dec 10 19:20:58 crc kubenswrapper[4828]: I1210 19:20:58.762740 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:58 crc kubenswrapper[4828]: I1210 19:20:58.762961 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:20:58 crc kubenswrapper[4828]: I1210 19:20:58.794139 4828 generic.go:334] "Generic (PLEG): container finished" podID="01a17451-e170-42ad-9082-d91eb11e237a" containerID="978bc8cc5cfbb1ea03b6dd0f9e5ed2ce4a7dd7ef8fac884740f815c131552c14" exitCode=0 Dec 10 19:20:58 crc kubenswrapper[4828]: I1210 19:20:58.794487 4828 generic.go:334] "Generic (PLEG): container finished" podID="01a17451-e170-42ad-9082-d91eb11e237a" containerID="f6176ab9eb78a7828354e6fc0062a0829cccd9dd942af8ae835ba5b60ee40f22" exitCode=2 Dec 10 19:20:58 crc kubenswrapper[4828]: I1210 19:20:58.794552 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01a17451-e170-42ad-9082-d91eb11e237a","Type":"ContainerDied","Data":"978bc8cc5cfbb1ea03b6dd0f9e5ed2ce4a7dd7ef8fac884740f815c131552c14"} Dec 10 19:20:58 crc kubenswrapper[4828]: I1210 19:20:58.794580 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01a17451-e170-42ad-9082-d91eb11e237a","Type":"ContainerDied","Data":"f6176ab9eb78a7828354e6fc0062a0829cccd9dd942af8ae835ba5b60ee40f22"} Dec 10 19:20:58 crc kubenswrapper[4828]: I1210 19:20:58.815999 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7b757664c-hkcjf" podStartSLOduration=4.815977941 podStartE2EDuration="4.815977941s" podCreationTimestamp="2025-12-10 19:20:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:58.795589764 +0000 UTC m=+1539.306200789" watchObservedRunningTime="2025-12-10 19:20:58.815977941 +0000 UTC m=+1539.326588946" Dec 10 19:20:58 crc kubenswrapper[4828]: I1210 19:20:58.834877 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"8f65c47b-68bb-4c64-b0e8-94507fb828e5","Type":"ContainerStarted","Data":"b441082c16a5ab19d22e3161d72fad78235f213d3bae6d0ee9c1be8d2e4858a5"} Dec 10 19:20:58 crc kubenswrapper[4828]: I1210 19:20:58.867763 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.867739814 podStartE2EDuration="5.867739814s" podCreationTimestamp="2025-12-10 19:20:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:20:58.865270009 +0000 UTC m=+1539.375881014" watchObservedRunningTime="2025-12-10 19:20:58.867739814 +0000 UTC m=+1539.378350829" Dec 10 19:20:59 crc kubenswrapper[4828]: I1210 19:20:59.030277 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 10 19:20:59 crc kubenswrapper[4828]: I1210 19:20:59.498509 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xmnbf" podUID="5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" containerName="registry-server" probeResult="failure" output=< Dec 10 19:20:59 crc kubenswrapper[4828]: timeout: failed to connect service ":50051" within 1s Dec 10 19:20:59 crc kubenswrapper[4828]: > Dec 10 19:20:59 crc kubenswrapper[4828]: I1210 19:20:59.876511 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" event={"ID":"41b1aa0c-723c-4f3f-9e52-26328c8ca954","Type":"ContainerStarted","Data":"532541bc74b35ce5d925fdafc6ed651ece85f11572169103a80f90b238b5d6e2"} Dec 10 19:20:59 crc kubenswrapper[4828]: I1210 19:20:59.877633 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:20:59 crc kubenswrapper[4828]: I1210 19:20:59.893961 4828 generic.go:334] "Generic (PLEG): container finished" podID="01a17451-e170-42ad-9082-d91eb11e237a" containerID="4b1e30bb14a3c5882d416795ef6afe70e2bae7117078dd92b6b8f1fd151fc78a" exitCode=0 Dec 10 19:20:59 crc kubenswrapper[4828]: I1210 19:20:59.893988 4828 generic.go:334] "Generic (PLEG): container finished" podID="01a17451-e170-42ad-9082-d91eb11e237a" containerID="3d7b46df6fba141148393519f246fb75354b03e9874c1e63610825d8b173d618" exitCode=0 Dec 10 19:20:59 crc kubenswrapper[4828]: I1210 19:20:59.895002 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01a17451-e170-42ad-9082-d91eb11e237a","Type":"ContainerDied","Data":"4b1e30bb14a3c5882d416795ef6afe70e2bae7117078dd92b6b8f1fd151fc78a"} Dec 10 19:20:59 crc kubenswrapper[4828]: I1210 19:20:59.895033 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01a17451-e170-42ad-9082-d91eb11e237a","Type":"ContainerDied","Data":"3d7b46df6fba141148393519f246fb75354b03e9874c1e63610825d8b173d618"} Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.072074 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" podStartSLOduration=5.072051799 podStartE2EDuration="5.072051799s" podCreationTimestamp="2025-12-10 19:20:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:21:00.030901125 +0000 UTC m=+1540.541512130" watchObservedRunningTime="2025-12-10 19:21:00.072051799 +0000 UTC m=+1540.582662804" Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.656496 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.695663 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-sg-core-conf-yaml\") pod \"01a17451-e170-42ad-9082-d91eb11e237a\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.695706 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01a17451-e170-42ad-9082-d91eb11e237a-log-httpd\") pod \"01a17451-e170-42ad-9082-d91eb11e237a\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.695731 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6hb7\" (UniqueName: \"kubernetes.io/projected/01a17451-e170-42ad-9082-d91eb11e237a-kube-api-access-k6hb7\") pod \"01a17451-e170-42ad-9082-d91eb11e237a\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.695756 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-scripts\") pod \"01a17451-e170-42ad-9082-d91eb11e237a\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.695794 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-config-data\") pod \"01a17451-e170-42ad-9082-d91eb11e237a\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.695861 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-combined-ca-bundle\") pod \"01a17451-e170-42ad-9082-d91eb11e237a\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.695916 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01a17451-e170-42ad-9082-d91eb11e237a-run-httpd\") pod \"01a17451-e170-42ad-9082-d91eb11e237a\" (UID: \"01a17451-e170-42ad-9082-d91eb11e237a\") " Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.696486 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01a17451-e170-42ad-9082-d91eb11e237a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "01a17451-e170-42ad-9082-d91eb11e237a" (UID: "01a17451-e170-42ad-9082-d91eb11e237a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.696728 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01a17451-e170-42ad-9082-d91eb11e237a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "01a17451-e170-42ad-9082-d91eb11e237a" (UID: "01a17451-e170-42ad-9082-d91eb11e237a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.702973 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01a17451-e170-42ad-9082-d91eb11e237a-kube-api-access-k6hb7" (OuterVolumeSpecName: "kube-api-access-k6hb7") pod "01a17451-e170-42ad-9082-d91eb11e237a" (UID: "01a17451-e170-42ad-9082-d91eb11e237a"). InnerVolumeSpecName "kube-api-access-k6hb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.734964 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-scripts" (OuterVolumeSpecName: "scripts") pod "01a17451-e170-42ad-9082-d91eb11e237a" (UID: "01a17451-e170-42ad-9082-d91eb11e237a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.764704 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "01a17451-e170-42ad-9082-d91eb11e237a" (UID: "01a17451-e170-42ad-9082-d91eb11e237a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.799261 4828 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.799303 4828 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01a17451-e170-42ad-9082-d91eb11e237a-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.799320 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6hb7\" (UniqueName: \"kubernetes.io/projected/01a17451-e170-42ad-9082-d91eb11e237a-kube-api-access-k6hb7\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.799335 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.799347 4828 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/01a17451-e170-42ad-9082-d91eb11e237a-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.858372 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01a17451-e170-42ad-9082-d91eb11e237a" (UID: "01a17451-e170-42ad-9082-d91eb11e237a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.905744 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.923166 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-config-data" (OuterVolumeSpecName: "config-data") pod "01a17451-e170-42ad-9082-d91eb11e237a" (UID: "01a17451-e170-42ad-9082-d91eb11e237a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.942057 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"01a17451-e170-42ad-9082-d91eb11e237a","Type":"ContainerDied","Data":"f37198571727cf7744f0fee52aa280d9a195a7dc1d503360bf2ed01ef2567f53"} Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.942132 4828 scope.go:117] "RemoveContainer" containerID="978bc8cc5cfbb1ea03b6dd0f9e5ed2ce4a7dd7ef8fac884740f815c131552c14" Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.942167 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:21:00 crc kubenswrapper[4828]: I1210 19:21:00.996159 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.008499 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01a17451-e170-42ad-9082-d91eb11e237a-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.019686 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.031010 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:21:01 crc kubenswrapper[4828]: E1210 19:21:01.031619 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a17451-e170-42ad-9082-d91eb11e237a" containerName="ceilometer-central-agent" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.031648 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a17451-e170-42ad-9082-d91eb11e237a" containerName="ceilometer-central-agent" Dec 10 19:21:01 crc kubenswrapper[4828]: E1210 19:21:01.031698 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a17451-e170-42ad-9082-d91eb11e237a" containerName="sg-core" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.031708 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a17451-e170-42ad-9082-d91eb11e237a" containerName="sg-core" Dec 10 19:21:01 crc kubenswrapper[4828]: E1210 19:21:01.031741 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a17451-e170-42ad-9082-d91eb11e237a" containerName="proxy-httpd" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.031752 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a17451-e170-42ad-9082-d91eb11e237a" containerName="proxy-httpd" Dec 10 19:21:01 crc kubenswrapper[4828]: E1210 19:21:01.031771 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01a17451-e170-42ad-9082-d91eb11e237a" containerName="ceilometer-notification-agent" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.031780 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="01a17451-e170-42ad-9082-d91eb11e237a" containerName="ceilometer-notification-agent" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.032095 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="01a17451-e170-42ad-9082-d91eb11e237a" containerName="sg-core" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.032122 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="01a17451-e170-42ad-9082-d91eb11e237a" containerName="proxy-httpd" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.032154 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="01a17451-e170-42ad-9082-d91eb11e237a" containerName="ceilometer-central-agent" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.032178 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="01a17451-e170-42ad-9082-d91eb11e237a" containerName="ceilometer-notification-agent" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.035515 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.041742 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.041913 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.068914 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.112879 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.112965 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-scripts\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.113050 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37649bc1-48ed-4cc3-9b7b-788469c5a989-run-httpd\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.113118 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37649bc1-48ed-4cc3-9b7b-788469c5a989-log-httpd\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.113257 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrv4t\" (UniqueName: \"kubernetes.io/projected/37649bc1-48ed-4cc3-9b7b-788469c5a989-kube-api-access-vrv4t\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.113289 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-config-data\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.113377 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.217132 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37649bc1-48ed-4cc3-9b7b-788469c5a989-log-httpd\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.217286 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrv4t\" (UniqueName: \"kubernetes.io/projected/37649bc1-48ed-4cc3-9b7b-788469c5a989-kube-api-access-vrv4t\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.217325 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-config-data\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.217385 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.217645 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37649bc1-48ed-4cc3-9b7b-788469c5a989-log-httpd\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.217664 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.217825 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-scripts\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.217969 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37649bc1-48ed-4cc3-9b7b-788469c5a989-run-httpd\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.218562 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37649bc1-48ed-4cc3-9b7b-788469c5a989-run-httpd\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.223352 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.224196 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-scripts\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.224454 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.227663 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-config-data\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.235717 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrv4t\" (UniqueName: \"kubernetes.io/projected/37649bc1-48ed-4cc3-9b7b-788469c5a989-kube-api-access-vrv4t\") pod \"ceilometer-0\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.369298 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:21:01 crc kubenswrapper[4828]: I1210 19:21:01.801117 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01a17451-e170-42ad-9082-d91eb11e237a" path="/var/lib/kubelet/pods/01a17451-e170-42ad-9082-d91eb11e237a/volumes" Dec 10 19:21:02 crc kubenswrapper[4828]: I1210 19:21:02.680213 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:21:02 crc kubenswrapper[4828]: I1210 19:21:02.680766 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c5296458-e976-4d87-a73f-3fb10cddd3d0" containerName="glance-log" containerID="cri-o://4655737d2ef0ac6fce17b004ea4fb37aeb4e7c011d3fb75feae97da21930b36a" gracePeriod=30 Dec 10 19:21:02 crc kubenswrapper[4828]: I1210 19:21:02.680921 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c5296458-e976-4d87-a73f-3fb10cddd3d0" containerName="glance-httpd" containerID="cri-o://9213c851c5e3a0adbad70dfa3fe4d86132ade02da751d4a97de33b2f0f9a6555" gracePeriod=30 Dec 10 19:21:02 crc kubenswrapper[4828]: I1210 19:21:02.983279 4828 generic.go:334] "Generic (PLEG): container finished" podID="c5296458-e976-4d87-a73f-3fb10cddd3d0" containerID="4655737d2ef0ac6fce17b004ea4fb37aeb4e7c011d3fb75feae97da21930b36a" exitCode=143 Dec 10 19:21:02 crc kubenswrapper[4828]: I1210 19:21:02.983491 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5296458-e976-4d87-a73f-3fb10cddd3d0","Type":"ContainerDied","Data":"4655737d2ef0ac6fce17b004ea4fb37aeb4e7c011d3fb75feae97da21930b36a"} Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.359675 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-d78dbcd66-dq2t9"] Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.361673 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.378921 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-566d76fd44-v6jl7"] Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.380981 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.395091 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-d78dbcd66-dq2t9"] Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.416584 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-6dfc7956d8-nq8qr"] Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.418121 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.451358 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-566d76fd44-v6jl7"] Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.505674 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-config-data\") pod \"heat-api-566d76fd44-v6jl7\" (UID: \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\") " pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.505893 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jl2q\" (UniqueName: \"kubernetes.io/projected/63fd1651-0e27-40c0-b5a4-38ab90a957ca-kube-api-access-2jl2q\") pod \"heat-api-566d76fd44-v6jl7\" (UID: \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\") " pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.505945 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-config-data-custom\") pod \"heat-api-566d76fd44-v6jl7\" (UID: \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\") " pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.506003 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-combined-ca-bundle\") pod \"heat-cfnapi-6dfc7956d8-nq8qr\" (UID: \"dbfc3632-aaef-44db-a52a-7403cae01ad4\") " pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.506115 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc9kc\" (UniqueName: \"kubernetes.io/projected/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-kube-api-access-fc9kc\") pod \"heat-engine-d78dbcd66-dq2t9\" (UID: \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\") " pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.506161 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-config-data\") pod \"heat-cfnapi-6dfc7956d8-nq8qr\" (UID: \"dbfc3632-aaef-44db-a52a-7403cae01ad4\") " pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.506195 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-combined-ca-bundle\") pod \"heat-api-566d76fd44-v6jl7\" (UID: \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\") " pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.506222 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-config-data-custom\") pod \"heat-engine-d78dbcd66-dq2t9\" (UID: \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\") " pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.506252 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-config-data\") pod \"heat-engine-d78dbcd66-dq2t9\" (UID: \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\") " pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.506295 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85rsj\" (UniqueName: \"kubernetes.io/projected/dbfc3632-aaef-44db-a52a-7403cae01ad4-kube-api-access-85rsj\") pod \"heat-cfnapi-6dfc7956d8-nq8qr\" (UID: \"dbfc3632-aaef-44db-a52a-7403cae01ad4\") " pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.506331 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-config-data-custom\") pod \"heat-cfnapi-6dfc7956d8-nq8qr\" (UID: \"dbfc3632-aaef-44db-a52a-7403cae01ad4\") " pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.506365 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-combined-ca-bundle\") pod \"heat-engine-d78dbcd66-dq2t9\" (UID: \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\") " pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.515587 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6dfc7956d8-nq8qr"] Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.608955 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-config-data-custom\") pod \"heat-cfnapi-6dfc7956d8-nq8qr\" (UID: \"dbfc3632-aaef-44db-a52a-7403cae01ad4\") " pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.609031 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-combined-ca-bundle\") pod \"heat-engine-d78dbcd66-dq2t9\" (UID: \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\") " pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.609099 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-config-data\") pod \"heat-api-566d76fd44-v6jl7\" (UID: \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\") " pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.609282 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jl2q\" (UniqueName: \"kubernetes.io/projected/63fd1651-0e27-40c0-b5a4-38ab90a957ca-kube-api-access-2jl2q\") pod \"heat-api-566d76fd44-v6jl7\" (UID: \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\") " pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.609309 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-config-data-custom\") pod \"heat-api-566d76fd44-v6jl7\" (UID: \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\") " pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.609370 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-combined-ca-bundle\") pod \"heat-cfnapi-6dfc7956d8-nq8qr\" (UID: \"dbfc3632-aaef-44db-a52a-7403cae01ad4\") " pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.609454 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc9kc\" (UniqueName: \"kubernetes.io/projected/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-kube-api-access-fc9kc\") pod \"heat-engine-d78dbcd66-dq2t9\" (UID: \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\") " pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.609499 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-config-data\") pod \"heat-cfnapi-6dfc7956d8-nq8qr\" (UID: \"dbfc3632-aaef-44db-a52a-7403cae01ad4\") " pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.609524 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-combined-ca-bundle\") pod \"heat-api-566d76fd44-v6jl7\" (UID: \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\") " pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.609543 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-config-data-custom\") pod \"heat-engine-d78dbcd66-dq2t9\" (UID: \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\") " pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.609570 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-config-data\") pod \"heat-engine-d78dbcd66-dq2t9\" (UID: \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\") " pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.609607 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85rsj\" (UniqueName: \"kubernetes.io/projected/dbfc3632-aaef-44db-a52a-7403cae01ad4-kube-api-access-85rsj\") pod \"heat-cfnapi-6dfc7956d8-nq8qr\" (UID: \"dbfc3632-aaef-44db-a52a-7403cae01ad4\") " pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.624551 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-combined-ca-bundle\") pod \"heat-engine-d78dbcd66-dq2t9\" (UID: \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\") " pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.624576 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-config-data-custom\") pod \"heat-engine-d78dbcd66-dq2t9\" (UID: \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\") " pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.627209 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-config-data\") pod \"heat-engine-d78dbcd66-dq2t9\" (UID: \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\") " pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.628591 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-config-data-custom\") pod \"heat-api-566d76fd44-v6jl7\" (UID: \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\") " pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.630685 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc9kc\" (UniqueName: \"kubernetes.io/projected/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-kube-api-access-fc9kc\") pod \"heat-engine-d78dbcd66-dq2t9\" (UID: \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\") " pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.632848 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-config-data-custom\") pod \"heat-cfnapi-6dfc7956d8-nq8qr\" (UID: \"dbfc3632-aaef-44db-a52a-7403cae01ad4\") " pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.637767 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-config-data\") pod \"heat-cfnapi-6dfc7956d8-nq8qr\" (UID: \"dbfc3632-aaef-44db-a52a-7403cae01ad4\") " pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.641269 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85rsj\" (UniqueName: \"kubernetes.io/projected/dbfc3632-aaef-44db-a52a-7403cae01ad4-kube-api-access-85rsj\") pod \"heat-cfnapi-6dfc7956d8-nq8qr\" (UID: \"dbfc3632-aaef-44db-a52a-7403cae01ad4\") " pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.641386 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-combined-ca-bundle\") pod \"heat-cfnapi-6dfc7956d8-nq8qr\" (UID: \"dbfc3632-aaef-44db-a52a-7403cae01ad4\") " pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.641691 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-combined-ca-bundle\") pod \"heat-api-566d76fd44-v6jl7\" (UID: \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\") " pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.649334 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jl2q\" (UniqueName: \"kubernetes.io/projected/63fd1651-0e27-40c0-b5a4-38ab90a957ca-kube-api-access-2jl2q\") pod \"heat-api-566d76fd44-v6jl7\" (UID: \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\") " pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.665715 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-config-data\") pod \"heat-api-566d76fd44-v6jl7\" (UID: \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\") " pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.697355 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.708116 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:03 crc kubenswrapper[4828]: I1210 19:21:03.760840 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:04 crc kubenswrapper[4828]: I1210 19:21:04.247024 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:21:04 crc kubenswrapper[4828]: I1210 19:21:04.247647 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="53369a6e-a6da-4d63-aca6-e1bec8ab047e" containerName="glance-log" containerID="cri-o://af613c0a15c1188568299015ab706de182e2534b39204ffc56468699e1878789" gracePeriod=30 Dec 10 19:21:04 crc kubenswrapper[4828]: I1210 19:21:04.247839 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="53369a6e-a6da-4d63-aca6-e1bec8ab047e" containerName="glance-httpd" containerID="cri-o://23700b3985da02baa67603281a49c24c027d44b3fb0df6c6db157900b864bded" gracePeriod=30 Dec 10 19:21:04 crc kubenswrapper[4828]: I1210 19:21:04.447205 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 10 19:21:04 crc kubenswrapper[4828]: I1210 19:21:04.790431 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:21:04 crc kubenswrapper[4828]: E1210 19:21:04.790702 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:21:04 crc kubenswrapper[4828]: I1210 19:21:04.822040 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:21:04 crc kubenswrapper[4828]: I1210 19:21:04.829624 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7b757664c-hkcjf" Dec 10 19:21:04 crc kubenswrapper[4828]: I1210 19:21:04.944567 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-jhdsp"] Dec 10 19:21:04 crc kubenswrapper[4828]: I1210 19:21:04.946960 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-jhdsp" Dec 10 19:21:04 crc kubenswrapper[4828]: I1210 19:21:04.977773 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-jhdsp"] Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.053783 4828 generic.go:334] "Generic (PLEG): container finished" podID="53369a6e-a6da-4d63-aca6-e1bec8ab047e" containerID="af613c0a15c1188568299015ab706de182e2534b39204ffc56468699e1878789" exitCode=143 Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.053836 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53369a6e-a6da-4d63-aca6-e1bec8ab047e","Type":"ContainerDied","Data":"af613c0a15c1188568299015ab706de182e2534b39204ffc56468699e1878789"} Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.056755 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lljp\" (UniqueName: \"kubernetes.io/projected/df859d5f-d811-4be3-82d8-8327c7cfb851-kube-api-access-7lljp\") pod \"nova-api-db-create-jhdsp\" (UID: \"df859d5f-d811-4be3-82d8-8327c7cfb851\") " pod="openstack/nova-api-db-create-jhdsp" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.056989 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df859d5f-d811-4be3-82d8-8327c7cfb851-operator-scripts\") pod \"nova-api-db-create-jhdsp\" (UID: \"df859d5f-d811-4be3-82d8-8327c7cfb851\") " pod="openstack/nova-api-db-create-jhdsp" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.058095 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-28cs6"] Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.059763 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-28cs6" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.080042 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-edd8-account-create-update-f87t8"] Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.082583 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-edd8-account-create-update-f87t8" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.085537 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.101767 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-28cs6"] Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.122872 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-edd8-account-create-update-f87t8"] Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.160246 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqzvj\" (UniqueName: \"kubernetes.io/projected/b544108c-9253-40f1-a5ed-8456173a3e68-kube-api-access-bqzvj\") pod \"nova-cell0-db-create-28cs6\" (UID: \"b544108c-9253-40f1-a5ed-8456173a3e68\") " pod="openstack/nova-cell0-db-create-28cs6" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.160568 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lljp\" (UniqueName: \"kubernetes.io/projected/df859d5f-d811-4be3-82d8-8327c7cfb851-kube-api-access-7lljp\") pod \"nova-api-db-create-jhdsp\" (UID: \"df859d5f-d811-4be3-82d8-8327c7cfb851\") " pod="openstack/nova-api-db-create-jhdsp" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.160751 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnb2k\" (UniqueName: \"kubernetes.io/projected/b19ce251-54b7-40f3-bbe0-6afb6064b307-kube-api-access-qnb2k\") pod \"nova-api-edd8-account-create-update-f87t8\" (UID: \"b19ce251-54b7-40f3-bbe0-6afb6064b307\") " pod="openstack/nova-api-edd8-account-create-update-f87t8" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.160968 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b544108c-9253-40f1-a5ed-8456173a3e68-operator-scripts\") pod \"nova-cell0-db-create-28cs6\" (UID: \"b544108c-9253-40f1-a5ed-8456173a3e68\") " pod="openstack/nova-cell0-db-create-28cs6" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.161004 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df859d5f-d811-4be3-82d8-8327c7cfb851-operator-scripts\") pod \"nova-api-db-create-jhdsp\" (UID: \"df859d5f-d811-4be3-82d8-8327c7cfb851\") " pod="openstack/nova-api-db-create-jhdsp" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.161059 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b19ce251-54b7-40f3-bbe0-6afb6064b307-operator-scripts\") pod \"nova-api-edd8-account-create-update-f87t8\" (UID: \"b19ce251-54b7-40f3-bbe0-6afb6064b307\") " pod="openstack/nova-api-edd8-account-create-update-f87t8" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.169591 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df859d5f-d811-4be3-82d8-8327c7cfb851-operator-scripts\") pod \"nova-api-db-create-jhdsp\" (UID: \"df859d5f-d811-4be3-82d8-8327c7cfb851\") " pod="openstack/nova-api-db-create-jhdsp" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.234820 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lljp\" (UniqueName: \"kubernetes.io/projected/df859d5f-d811-4be3-82d8-8327c7cfb851-kube-api-access-7lljp\") pod \"nova-api-db-create-jhdsp\" (UID: \"df859d5f-d811-4be3-82d8-8327c7cfb851\") " pod="openstack/nova-api-db-create-jhdsp" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.272888 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-5tqrl"] Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.273652 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnb2k\" (UniqueName: \"kubernetes.io/projected/b19ce251-54b7-40f3-bbe0-6afb6064b307-kube-api-access-qnb2k\") pod \"nova-api-edd8-account-create-update-f87t8\" (UID: \"b19ce251-54b7-40f3-bbe0-6afb6064b307\") " pod="openstack/nova-api-edd8-account-create-update-f87t8" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.273736 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b544108c-9253-40f1-a5ed-8456173a3e68-operator-scripts\") pod \"nova-cell0-db-create-28cs6\" (UID: \"b544108c-9253-40f1-a5ed-8456173a3e68\") " pod="openstack/nova-cell0-db-create-28cs6" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.273770 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b19ce251-54b7-40f3-bbe0-6afb6064b307-operator-scripts\") pod \"nova-api-edd8-account-create-update-f87t8\" (UID: \"b19ce251-54b7-40f3-bbe0-6afb6064b307\") " pod="openstack/nova-api-edd8-account-create-update-f87t8" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.273868 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqzvj\" (UniqueName: \"kubernetes.io/projected/b544108c-9253-40f1-a5ed-8456173a3e68-kube-api-access-bqzvj\") pod \"nova-cell0-db-create-28cs6\" (UID: \"b544108c-9253-40f1-a5ed-8456173a3e68\") " pod="openstack/nova-cell0-db-create-28cs6" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.274970 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5tqrl" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.275008 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b544108c-9253-40f1-a5ed-8456173a3e68-operator-scripts\") pod \"nova-cell0-db-create-28cs6\" (UID: \"b544108c-9253-40f1-a5ed-8456173a3e68\") " pod="openstack/nova-cell0-db-create-28cs6" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.275504 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b19ce251-54b7-40f3-bbe0-6afb6064b307-operator-scripts\") pod \"nova-api-edd8-account-create-update-f87t8\" (UID: \"b19ce251-54b7-40f3-bbe0-6afb6064b307\") " pod="openstack/nova-api-edd8-account-create-update-f87t8" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.302046 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5tqrl"] Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.304247 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-jhdsp" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.323718 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnb2k\" (UniqueName: \"kubernetes.io/projected/b19ce251-54b7-40f3-bbe0-6afb6064b307-kube-api-access-qnb2k\") pod \"nova-api-edd8-account-create-update-f87t8\" (UID: \"b19ce251-54b7-40f3-bbe0-6afb6064b307\") " pod="openstack/nova-api-edd8-account-create-update-f87t8" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.344482 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqzvj\" (UniqueName: \"kubernetes.io/projected/b544108c-9253-40f1-a5ed-8456173a3e68-kube-api-access-bqzvj\") pod \"nova-cell0-db-create-28cs6\" (UID: \"b544108c-9253-40f1-a5ed-8456173a3e68\") " pod="openstack/nova-cell0-db-create-28cs6" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.344562 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-e96e-account-create-update-8kssp"] Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.346266 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e96e-account-create-update-8kssp" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.368681 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.378397 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft78c\" (UniqueName: \"kubernetes.io/projected/d24aa875-f0bd-4e7d-a985-c633a246fb30-kube-api-access-ft78c\") pod \"nova-cell1-db-create-5tqrl\" (UID: \"d24aa875-f0bd-4e7d-a985-c633a246fb30\") " pod="openstack/nova-cell1-db-create-5tqrl" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.378763 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d24aa875-f0bd-4e7d-a985-c633a246fb30-operator-scripts\") pod \"nova-cell1-db-create-5tqrl\" (UID: \"d24aa875-f0bd-4e7d-a985-c633a246fb30\") " pod="openstack/nova-cell1-db-create-5tqrl" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.379367 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-28cs6" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.438398 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-edd8-account-create-update-f87t8" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.455078 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-e96e-account-create-update-8kssp"] Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.481303 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f5c519b-682a-430f-b1ed-ee93f6b7c9f2-operator-scripts\") pod \"nova-cell0-e96e-account-create-update-8kssp\" (UID: \"1f5c519b-682a-430f-b1ed-ee93f6b7c9f2\") " pod="openstack/nova-cell0-e96e-account-create-update-8kssp" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.481473 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft78c\" (UniqueName: \"kubernetes.io/projected/d24aa875-f0bd-4e7d-a985-c633a246fb30-kube-api-access-ft78c\") pod \"nova-cell1-db-create-5tqrl\" (UID: \"d24aa875-f0bd-4e7d-a985-c633a246fb30\") " pod="openstack/nova-cell1-db-create-5tqrl" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.481636 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zjkp\" (UniqueName: \"kubernetes.io/projected/1f5c519b-682a-430f-b1ed-ee93f6b7c9f2-kube-api-access-2zjkp\") pod \"nova-cell0-e96e-account-create-update-8kssp\" (UID: \"1f5c519b-682a-430f-b1ed-ee93f6b7c9f2\") " pod="openstack/nova-cell0-e96e-account-create-update-8kssp" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.481688 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d24aa875-f0bd-4e7d-a985-c633a246fb30-operator-scripts\") pod \"nova-cell1-db-create-5tqrl\" (UID: \"d24aa875-f0bd-4e7d-a985-c633a246fb30\") " pod="openstack/nova-cell1-db-create-5tqrl" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.483123 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d24aa875-f0bd-4e7d-a985-c633a246fb30-operator-scripts\") pod \"nova-cell1-db-create-5tqrl\" (UID: \"d24aa875-f0bd-4e7d-a985-c633a246fb30\") " pod="openstack/nova-cell1-db-create-5tqrl" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.512914 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-c67d-account-create-update-d6zs6"] Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.520103 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c67d-account-create-update-d6zs6" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.529349 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.539137 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-c67d-account-create-update-d6zs6"] Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.574612 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft78c\" (UniqueName: \"kubernetes.io/projected/d24aa875-f0bd-4e7d-a985-c633a246fb30-kube-api-access-ft78c\") pod \"nova-cell1-db-create-5tqrl\" (UID: \"d24aa875-f0bd-4e7d-a985-c633a246fb30\") " pod="openstack/nova-cell1-db-create-5tqrl" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.584083 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f5c519b-682a-430f-b1ed-ee93f6b7c9f2-operator-scripts\") pod \"nova-cell0-e96e-account-create-update-8kssp\" (UID: \"1f5c519b-682a-430f-b1ed-ee93f6b7c9f2\") " pod="openstack/nova-cell0-e96e-account-create-update-8kssp" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.584222 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47be5ead-d925-4058-a390-88426775ebfa-operator-scripts\") pod \"nova-cell1-c67d-account-create-update-d6zs6\" (UID: \"47be5ead-d925-4058-a390-88426775ebfa\") " pod="openstack/nova-cell1-c67d-account-create-update-d6zs6" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.584296 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zjkp\" (UniqueName: \"kubernetes.io/projected/1f5c519b-682a-430f-b1ed-ee93f6b7c9f2-kube-api-access-2zjkp\") pod \"nova-cell0-e96e-account-create-update-8kssp\" (UID: \"1f5c519b-682a-430f-b1ed-ee93f6b7c9f2\") " pod="openstack/nova-cell0-e96e-account-create-update-8kssp" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.584355 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh6vf\" (UniqueName: \"kubernetes.io/projected/47be5ead-d925-4058-a390-88426775ebfa-kube-api-access-lh6vf\") pod \"nova-cell1-c67d-account-create-update-d6zs6\" (UID: \"47be5ead-d925-4058-a390-88426775ebfa\") " pod="openstack/nova-cell1-c67d-account-create-update-d6zs6" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.586077 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f5c519b-682a-430f-b1ed-ee93f6b7c9f2-operator-scripts\") pod \"nova-cell0-e96e-account-create-update-8kssp\" (UID: \"1f5c519b-682a-430f-b1ed-ee93f6b7c9f2\") " pod="openstack/nova-cell0-e96e-account-create-update-8kssp" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.600570 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zjkp\" (UniqueName: \"kubernetes.io/projected/1f5c519b-682a-430f-b1ed-ee93f6b7c9f2-kube-api-access-2zjkp\") pod \"nova-cell0-e96e-account-create-update-8kssp\" (UID: \"1f5c519b-682a-430f-b1ed-ee93f6b7c9f2\") " pod="openstack/nova-cell0-e96e-account-create-update-8kssp" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.687047 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47be5ead-d925-4058-a390-88426775ebfa-operator-scripts\") pod \"nova-cell1-c67d-account-create-update-d6zs6\" (UID: \"47be5ead-d925-4058-a390-88426775ebfa\") " pod="openstack/nova-cell1-c67d-account-create-update-d6zs6" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.688008 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47be5ead-d925-4058-a390-88426775ebfa-operator-scripts\") pod \"nova-cell1-c67d-account-create-update-d6zs6\" (UID: \"47be5ead-d925-4058-a390-88426775ebfa\") " pod="openstack/nova-cell1-c67d-account-create-update-d6zs6" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.691096 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh6vf\" (UniqueName: \"kubernetes.io/projected/47be5ead-d925-4058-a390-88426775ebfa-kube-api-access-lh6vf\") pod \"nova-cell1-c67d-account-create-update-d6zs6\" (UID: \"47be5ead-d925-4058-a390-88426775ebfa\") " pod="openstack/nova-cell1-c67d-account-create-update-d6zs6" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.707581 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh6vf\" (UniqueName: \"kubernetes.io/projected/47be5ead-d925-4058-a390-88426775ebfa-kube-api-access-lh6vf\") pod \"nova-cell1-c67d-account-create-update-d6zs6\" (UID: \"47be5ead-d925-4058-a390-88426775ebfa\") " pod="openstack/nova-cell1-c67d-account-create-update-d6zs6" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.718432 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.756606 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5tqrl" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.780500 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e96e-account-create-update-8kssp" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.847715 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c67d-account-create-update-d6zs6" Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.875358 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-s5bns"] Dec 10 19:21:05 crc kubenswrapper[4828]: I1210 19:21:05.875647 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-s5bns" podUID="fd2aa65d-1eae-421a-9198-684337484b92" containerName="dnsmasq-dns" containerID="cri-o://eba8b69562e3c40b6f7239cf498dc8bc63213a06fbf4e6be415111e00411e688" gracePeriod=10 Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.071511 4828 generic.go:334] "Generic (PLEG): container finished" podID="fd2aa65d-1eae-421a-9198-684337484b92" containerID="eba8b69562e3c40b6f7239cf498dc8bc63213a06fbf4e6be415111e00411e688" exitCode=0 Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.071568 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-s5bns" event={"ID":"fd2aa65d-1eae-421a-9198-684337484b92","Type":"ContainerDied","Data":"eba8b69562e3c40b6f7239cf498dc8bc63213a06fbf4e6be415111e00411e688"} Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.088877 4828 generic.go:334] "Generic (PLEG): container finished" podID="c5296458-e976-4d87-a73f-3fb10cddd3d0" containerID="9213c851c5e3a0adbad70dfa3fe4d86132ade02da751d4a97de33b2f0f9a6555" exitCode=0 Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.088926 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5296458-e976-4d87-a73f-3fb10cddd3d0","Type":"ContainerDied","Data":"9213c851c5e3a0adbad70dfa3fe4d86132ade02da751d4a97de33b2f0f9a6555"} Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.407127 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-77bd4457c-f5tqz"] Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.430458 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6f7b6d44ff-rhjwx"] Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.461833 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-59c4ffddfc-nf4zg"] Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.463866 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.471189 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.471424 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.479457 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-546dfdfc9f-s9ctn"] Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.483656 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.486685 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.486867 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.505160 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-59c4ffddfc-nf4zg"] Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.509548 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-config-data\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.509644 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-config-data-custom\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.509709 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-public-tls-certs\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.509735 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-config-data\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.509771 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-combined-ca-bundle\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.509845 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8v5r\" (UniqueName: \"kubernetes.io/projected/296faf68-6251-48d1-bc36-026a31c00393-kube-api-access-t8v5r\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.509877 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-public-tls-certs\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.509895 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-internal-tls-certs\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.509943 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-internal-tls-certs\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.509967 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-combined-ca-bundle\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.509996 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lptvz\" (UniqueName: \"kubernetes.io/projected/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-kube-api-access-lptvz\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.510033 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-config-data-custom\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.527274 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-546dfdfc9f-s9ctn"] Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.611700 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lptvz\" (UniqueName: \"kubernetes.io/projected/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-kube-api-access-lptvz\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.612112 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-config-data-custom\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.612224 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-config-data\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.612376 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-config-data-custom\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.612540 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-public-tls-certs\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.612649 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-config-data\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.612747 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-combined-ca-bundle\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.612970 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8v5r\" (UniqueName: \"kubernetes.io/projected/296faf68-6251-48d1-bc36-026a31c00393-kube-api-access-t8v5r\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.613098 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-public-tls-certs\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.613194 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-internal-tls-certs\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.613368 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-internal-tls-certs\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.613469 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-combined-ca-bundle\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.618124 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-combined-ca-bundle\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.620573 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-config-data\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.624120 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-internal-tls-certs\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.625508 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-public-tls-certs\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.626718 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-config-data-custom\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.627543 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-public-tls-certs\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.627854 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-combined-ca-bundle\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.628443 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-config-data-custom\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.638534 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-config-data\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.639477 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lptvz\" (UniqueName: \"kubernetes.io/projected/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-kube-api-access-lptvz\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.639906 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-internal-tls-certs\") pod \"heat-api-59c4ffddfc-nf4zg\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.660770 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8v5r\" (UniqueName: \"kubernetes.io/projected/296faf68-6251-48d1-bc36-026a31c00393-kube-api-access-t8v5r\") pod \"heat-cfnapi-546dfdfc9f-s9ctn\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.797465 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:06 crc kubenswrapper[4828]: I1210 19:21:06.836104 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:07 crc kubenswrapper[4828]: I1210 19:21:07.510590 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:21:07 crc kubenswrapper[4828]: I1210 19:21:07.513558 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:21:07 crc kubenswrapper[4828]: I1210 19:21:07.531499 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-f568c66f6-nt8g5" Dec 10 19:21:08 crc kubenswrapper[4828]: I1210 19:21:08.151419 4828 generic.go:334] "Generic (PLEG): container finished" podID="53369a6e-a6da-4d63-aca6-e1bec8ab047e" containerID="23700b3985da02baa67603281a49c24c027d44b3fb0df6c6db157900b864bded" exitCode=0 Dec 10 19:21:08 crc kubenswrapper[4828]: I1210 19:21:08.151544 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53369a6e-a6da-4d63-aca6-e1bec8ab047e","Type":"ContainerDied","Data":"23700b3985da02baa67603281a49c24c027d44b3fb0df6c6db157900b864bded"} Dec 10 19:21:08 crc kubenswrapper[4828]: I1210 19:21:08.485316 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:21:08 crc kubenswrapper[4828]: I1210 19:21:08.535657 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:21:08 crc kubenswrapper[4828]: I1210 19:21:08.726351 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xmnbf"] Dec 10 19:21:08 crc kubenswrapper[4828]: I1210 19:21:08.743709 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="c5296458-e976-4d87-a73f-3fb10cddd3d0" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.186:9292/healthcheck\": dial tcp 10.217.0.186:9292: connect: connection refused" Dec 10 19:21:08 crc kubenswrapper[4828]: I1210 19:21:08.743786 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="c5296458-e976-4d87-a73f-3fb10cddd3d0" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.186:9292/healthcheck\": dial tcp 10.217.0.186:9292: connect: connection refused" Dec 10 19:21:09 crc kubenswrapper[4828]: I1210 19:21:09.340875 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="53369a6e-a6da-4d63-aca6-e1bec8ab047e" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.187:9292/healthcheck\": dial tcp 10.217.0.187:9292: connect: connection refused" Dec 10 19:21:09 crc kubenswrapper[4828]: I1210 19:21:09.341454 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="53369a6e-a6da-4d63-aca6-e1bec8ab047e" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.187:9292/healthcheck\": dial tcp 10.217.0.187:9292: connect: connection refused" Dec 10 19:21:09 crc kubenswrapper[4828]: I1210 19:21:09.413111 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6578955fd5-s5bns" podUID="fd2aa65d-1eae-421a-9198-684337484b92" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.199:5353: connect: connection refused" Dec 10 19:21:10 crc kubenswrapper[4828]: I1210 19:21:10.176966 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xmnbf" podUID="5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" containerName="registry-server" containerID="cri-o://aba92744fea3340582f4b08edcdecf604363ec39c5ed811390e34f2ec576d930" gracePeriod=2 Dec 10 19:21:10 crc kubenswrapper[4828]: I1210 19:21:10.669198 4828 scope.go:117] "RemoveContainer" containerID="f6176ab9eb78a7828354e6fc0062a0829cccd9dd942af8ae835ba5b60ee40f22" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.209437 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.255474 4828 generic.go:334] "Generic (PLEG): container finished" podID="5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" containerID="aba92744fea3340582f4b08edcdecf604363ec39c5ed811390e34f2ec576d930" exitCode=0 Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.255753 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xmnbf" event={"ID":"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a","Type":"ContainerDied","Data":"aba92744fea3340582f4b08edcdecf604363ec39c5ed811390e34f2ec576d930"} Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.255782 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xmnbf" event={"ID":"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a","Type":"ContainerDied","Data":"9a190a12c747d6f01a8236f0a84d43d5ea1b992eec960a0364ad5cea3401f1d4"} Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.255859 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.267358 4828 scope.go:117] "RemoveContainer" containerID="4b1e30bb14a3c5882d416795ef6afe70e2bae7117078dd92b6b8f1fd151fc78a" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.354398 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-catalog-content\") pod \"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a\" (UID: \"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a\") " Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.354501 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p97dz\" (UniqueName: \"kubernetes.io/projected/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-kube-api-access-p97dz\") pod \"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a\" (UID: \"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a\") " Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.354764 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-utilities\") pod \"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a\" (UID: \"5f9f6cc8-f9aa-478c-9694-bcc05679ba3a\") " Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.355251 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-utilities" (OuterVolumeSpecName: "utilities") pod "5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" (UID: "5f9f6cc8-f9aa-478c-9694-bcc05679ba3a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.355467 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.363623 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-kube-api-access-p97dz" (OuterVolumeSpecName: "kube-api-access-p97dz") pod "5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" (UID: "5f9f6cc8-f9aa-478c-9694-bcc05679ba3a"). InnerVolumeSpecName "kube-api-access-p97dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.461868 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p97dz\" (UniqueName: \"kubernetes.io/projected/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-kube-api-access-p97dz\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.464299 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.515027 4828 scope.go:117] "RemoveContainer" containerID="3d7b46df6fba141148393519f246fb75354b03e9874c1e63610825d8b173d618" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.567890 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lb9hw\" (UniqueName: \"kubernetes.io/projected/fd2aa65d-1eae-421a-9198-684337484b92-kube-api-access-lb9hw\") pod \"fd2aa65d-1eae-421a-9198-684337484b92\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.568003 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-dns-svc\") pod \"fd2aa65d-1eae-421a-9198-684337484b92\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.568073 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-ovsdbserver-sb\") pod \"fd2aa65d-1eae-421a-9198-684337484b92\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.568316 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-ovsdbserver-nb\") pod \"fd2aa65d-1eae-421a-9198-684337484b92\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.568399 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-config\") pod \"fd2aa65d-1eae-421a-9198-684337484b92\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.570193 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" (UID: "5f9f6cc8-f9aa-478c-9694-bcc05679ba3a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.570609 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.616875 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd2aa65d-1eae-421a-9198-684337484b92-kube-api-access-lb9hw" (OuterVolumeSpecName: "kube-api-access-lb9hw") pod "fd2aa65d-1eae-421a-9198-684337484b92" (UID: "fd2aa65d-1eae-421a-9198-684337484b92"). InnerVolumeSpecName "kube-api-access-lb9hw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.672479 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-dns-swift-storage-0\") pod \"fd2aa65d-1eae-421a-9198-684337484b92\" (UID: \"fd2aa65d-1eae-421a-9198-684337484b92\") " Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.674158 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lb9hw\" (UniqueName: \"kubernetes.io/projected/fd2aa65d-1eae-421a-9198-684337484b92-kube-api-access-lb9hw\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.742454 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-config" (OuterVolumeSpecName: "config") pod "fd2aa65d-1eae-421a-9198-684337484b92" (UID: "fd2aa65d-1eae-421a-9198-684337484b92"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.746215 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fd2aa65d-1eae-421a-9198-684337484b92" (UID: "fd2aa65d-1eae-421a-9198-684337484b92"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.758314 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fd2aa65d-1eae-421a-9198-684337484b92" (UID: "fd2aa65d-1eae-421a-9198-684337484b92"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.762046 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fd2aa65d-1eae-421a-9198-684337484b92" (UID: "fd2aa65d-1eae-421a-9198-684337484b92"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.776301 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.776333 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.776342 4828 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.776350 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.822493 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fd2aa65d-1eae-421a-9198-684337484b92" (UID: "fd2aa65d-1eae-421a-9198-684337484b92"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.880781 4828 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd2aa65d-1eae-421a-9198-684337484b92-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.918851 4828 scope.go:117] "RemoveContainer" containerID="aba92744fea3340582f4b08edcdecf604363ec39c5ed811390e34f2ec576d930" Dec 10 19:21:11 crc kubenswrapper[4828]: I1210 19:21:11.987191 4828 scope.go:117] "RemoveContainer" containerID="d1b567629ae86099880aff767da3052efee6a1b9fb6c3120df856e309ce2a1e5" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.017334 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.065030 4828 scope.go:117] "RemoveContainer" containerID="5f1f7213d7475e6ee288c2556ef8fb5123714ba9df2618833f23317f99b506e5" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.095903 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-config-data\") pod \"c5296458-e976-4d87-a73f-3fb10cddd3d0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.095991 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-public-tls-certs\") pod \"c5296458-e976-4d87-a73f-3fb10cddd3d0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.096305 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-combined-ca-bundle\") pod \"c5296458-e976-4d87-a73f-3fb10cddd3d0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.096392 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5296458-e976-4d87-a73f-3fb10cddd3d0-httpd-run\") pod \"c5296458-e976-4d87-a73f-3fb10cddd3d0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.096451 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-scripts\") pod \"c5296458-e976-4d87-a73f-3fb10cddd3d0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.097756 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5296458-e976-4d87-a73f-3fb10cddd3d0-logs\") pod \"c5296458-e976-4d87-a73f-3fb10cddd3d0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.098330 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"c5296458-e976-4d87-a73f-3fb10cddd3d0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.098428 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9llcc\" (UniqueName: \"kubernetes.io/projected/c5296458-e976-4d87-a73f-3fb10cddd3d0-kube-api-access-9llcc\") pod \"c5296458-e976-4d87-a73f-3fb10cddd3d0\" (UID: \"c5296458-e976-4d87-a73f-3fb10cddd3d0\") " Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.100483 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5296458-e976-4d87-a73f-3fb10cddd3d0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c5296458-e976-4d87-a73f-3fb10cddd3d0" (UID: "c5296458-e976-4d87-a73f-3fb10cddd3d0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.100981 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5296458-e976-4d87-a73f-3fb10cddd3d0-logs" (OuterVolumeSpecName: "logs") pod "c5296458-e976-4d87-a73f-3fb10cddd3d0" (UID: "c5296458-e976-4d87-a73f-3fb10cddd3d0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.117470 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "c5296458-e976-4d87-a73f-3fb10cddd3d0" (UID: "c5296458-e976-4d87-a73f-3fb10cddd3d0"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.117923 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5296458-e976-4d87-a73f-3fb10cddd3d0-kube-api-access-9llcc" (OuterVolumeSpecName: "kube-api-access-9llcc") pod "c5296458-e976-4d87-a73f-3fb10cddd3d0" (UID: "c5296458-e976-4d87-a73f-3fb10cddd3d0"). InnerVolumeSpecName "kube-api-access-9llcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.118507 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-scripts" (OuterVolumeSpecName: "scripts") pod "c5296458-e976-4d87-a73f-3fb10cddd3d0" (UID: "c5296458-e976-4d87-a73f-3fb10cddd3d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.138936 4828 scope.go:117] "RemoveContainer" containerID="aba92744fea3340582f4b08edcdecf604363ec39c5ed811390e34f2ec576d930" Dec 10 19:21:12 crc kubenswrapper[4828]: E1210 19:21:12.139791 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aba92744fea3340582f4b08edcdecf604363ec39c5ed811390e34f2ec576d930\": container with ID starting with aba92744fea3340582f4b08edcdecf604363ec39c5ed811390e34f2ec576d930 not found: ID does not exist" containerID="aba92744fea3340582f4b08edcdecf604363ec39c5ed811390e34f2ec576d930" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.139849 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aba92744fea3340582f4b08edcdecf604363ec39c5ed811390e34f2ec576d930"} err="failed to get container status \"aba92744fea3340582f4b08edcdecf604363ec39c5ed811390e34f2ec576d930\": rpc error: code = NotFound desc = could not find container \"aba92744fea3340582f4b08edcdecf604363ec39c5ed811390e34f2ec576d930\": container with ID starting with aba92744fea3340582f4b08edcdecf604363ec39c5ed811390e34f2ec576d930 not found: ID does not exist" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.139877 4828 scope.go:117] "RemoveContainer" containerID="d1b567629ae86099880aff767da3052efee6a1b9fb6c3120df856e309ce2a1e5" Dec 10 19:21:12 crc kubenswrapper[4828]: E1210 19:21:12.140317 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1b567629ae86099880aff767da3052efee6a1b9fb6c3120df856e309ce2a1e5\": container with ID starting with d1b567629ae86099880aff767da3052efee6a1b9fb6c3120df856e309ce2a1e5 not found: ID does not exist" containerID="d1b567629ae86099880aff767da3052efee6a1b9fb6c3120df856e309ce2a1e5" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.140352 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1b567629ae86099880aff767da3052efee6a1b9fb6c3120df856e309ce2a1e5"} err="failed to get container status \"d1b567629ae86099880aff767da3052efee6a1b9fb6c3120df856e309ce2a1e5\": rpc error: code = NotFound desc = could not find container \"d1b567629ae86099880aff767da3052efee6a1b9fb6c3120df856e309ce2a1e5\": container with ID starting with d1b567629ae86099880aff767da3052efee6a1b9fb6c3120df856e309ce2a1e5 not found: ID does not exist" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.140373 4828 scope.go:117] "RemoveContainer" containerID="5f1f7213d7475e6ee288c2556ef8fb5123714ba9df2618833f23317f99b506e5" Dec 10 19:21:12 crc kubenswrapper[4828]: E1210 19:21:12.140684 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f1f7213d7475e6ee288c2556ef8fb5123714ba9df2618833f23317f99b506e5\": container with ID starting with 5f1f7213d7475e6ee288c2556ef8fb5123714ba9df2618833f23317f99b506e5 not found: ID does not exist" containerID="5f1f7213d7475e6ee288c2556ef8fb5123714ba9df2618833f23317f99b506e5" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.140739 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f1f7213d7475e6ee288c2556ef8fb5123714ba9df2618833f23317f99b506e5"} err="failed to get container status \"5f1f7213d7475e6ee288c2556ef8fb5123714ba9df2618833f23317f99b506e5\": rpc error: code = NotFound desc = could not find container \"5f1f7213d7475e6ee288c2556ef8fb5123714ba9df2618833f23317f99b506e5\": container with ID starting with 5f1f7213d7475e6ee288c2556ef8fb5123714ba9df2618833f23317f99b506e5 not found: ID does not exist" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.202340 4828 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c5296458-e976-4d87-a73f-3fb10cddd3d0-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.203742 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.203784 4828 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5296458-e976-4d87-a73f-3fb10cddd3d0-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.203844 4828 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.203862 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9llcc\" (UniqueName: \"kubernetes.io/projected/c5296458-e976-4d87-a73f-3fb10cddd3d0-kube-api-access-9llcc\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.283956 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-s5bns" event={"ID":"fd2aa65d-1eae-421a-9198-684337484b92","Type":"ContainerDied","Data":"bae93beb40f5418760d4f13a83c452e9fb4a37adf58d5674257fae52b0ea95a0"} Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.284019 4828 scope.go:117] "RemoveContainer" containerID="eba8b69562e3c40b6f7239cf498dc8bc63213a06fbf4e6be415111e00411e688" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.284140 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-s5bns" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.344310 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c5296458-e976-4d87-a73f-3fb10cddd3d0","Type":"ContainerDied","Data":"4a451e3223768c991f379f410c411c08c1783117a52e4d37b5e4cbd9f7adc27e"} Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.344408 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.386898 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-s5bns"] Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.398437 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-s5bns"] Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.411266 4828 scope.go:117] "RemoveContainer" containerID="3f279e3ec9b46a37bad62bca048f38da14b00d3ddfe4cf78080f569f2e589218" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.530030 4828 scope.go:117] "RemoveContainer" containerID="9213c851c5e3a0adbad70dfa3fe4d86132ade02da751d4a97de33b2f0f9a6555" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.540496 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6dfc7956d8-nq8qr"] Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.546203 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5296458-e976-4d87-a73f-3fb10cddd3d0" (UID: "c5296458-e976-4d87-a73f-3fb10cddd3d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:12 crc kubenswrapper[4828]: W1210 19:21:12.554198 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbfc3632_aaef_44db_a52a_7403cae01ad4.slice/crio-88ccd8b0d00966fb6e4dadab1cc9c578a434c80e53a522a22b1f66ab116dfbe2 WatchSource:0}: Error finding container 88ccd8b0d00966fb6e4dadab1cc9c578a434c80e53a522a22b1f66ab116dfbe2: Status 404 returned error can't find the container with id 88ccd8b0d00966fb6e4dadab1cc9c578a434c80e53a522a22b1f66ab116dfbe2 Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.563272 4828 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.619973 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.623499 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.623534 4828 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.668645 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c5296458-e976-4d87-a73f-3fb10cddd3d0" (UID: "c5296458-e976-4d87-a73f-3fb10cddd3d0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.672747 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-config-data" (OuterVolumeSpecName: "config-data") pod "c5296458-e976-4d87-a73f-3fb10cddd3d0" (UID: "c5296458-e976-4d87-a73f-3fb10cddd3d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.727056 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.727095 4828 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5296458-e976-4d87-a73f-3fb10cddd3d0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.813696 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-jhdsp"] Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.824172 4828 scope.go:117] "RemoveContainer" containerID="4655737d2ef0ac6fce17b004ea4fb37aeb4e7c011d3fb75feae97da21930b36a" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.887312 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.931592 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-internal-tls-certs\") pod \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.931691 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-scripts\") pod \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.931775 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-config-data\") pod \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.931815 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75ppq\" (UniqueName: \"kubernetes.io/projected/53369a6e-a6da-4d63-aca6-e1bec8ab047e-kube-api-access-75ppq\") pod \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.931957 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53369a6e-a6da-4d63-aca6-e1bec8ab047e-logs\") pod \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.932000 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-combined-ca-bundle\") pod \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.932026 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.932055 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53369a6e-a6da-4d63-aca6-e1bec8ab047e-httpd-run\") pod \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\" (UID: \"53369a6e-a6da-4d63-aca6-e1bec8ab047e\") " Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.932893 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53369a6e-a6da-4d63-aca6-e1bec8ab047e-logs" (OuterVolumeSpecName: "logs") pod "53369a6e-a6da-4d63-aca6-e1bec8ab047e" (UID: "53369a6e-a6da-4d63-aca6-e1bec8ab047e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.932984 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53369a6e-a6da-4d63-aca6-e1bec8ab047e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "53369a6e-a6da-4d63-aca6-e1bec8ab047e" (UID: "53369a6e-a6da-4d63-aca6-e1bec8ab047e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.994073 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "53369a6e-a6da-4d63-aca6-e1bec8ab047e" (UID: "53369a6e-a6da-4d63-aca6-e1bec8ab047e"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.995066 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-scripts" (OuterVolumeSpecName: "scripts") pod "53369a6e-a6da-4d63-aca6-e1bec8ab047e" (UID: "53369a6e-a6da-4d63-aca6-e1bec8ab047e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:12 crc kubenswrapper[4828]: I1210 19:21:12.996068 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53369a6e-a6da-4d63-aca6-e1bec8ab047e-kube-api-access-75ppq" (OuterVolumeSpecName: "kube-api-access-75ppq") pod "53369a6e-a6da-4d63-aca6-e1bec8ab047e" (UID: "53369a6e-a6da-4d63-aca6-e1bec8ab047e"). InnerVolumeSpecName "kube-api-access-75ppq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.042446 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75ppq\" (UniqueName: \"kubernetes.io/projected/53369a6e-a6da-4d63-aca6-e1bec8ab047e-kube-api-access-75ppq\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.042486 4828 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53369a6e-a6da-4d63-aca6-e1bec8ab047e-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.042513 4828 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.042527 4828 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53369a6e-a6da-4d63-aca6-e1bec8ab047e-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.042538 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.054069 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-config-data" (OuterVolumeSpecName: "config-data") pod "53369a6e-a6da-4d63-aca6-e1bec8ab047e" (UID: "53369a6e-a6da-4d63-aca6-e1bec8ab047e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.056118 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.076088 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.099753 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:21:13 crc kubenswrapper[4828]: E1210 19:21:13.100248 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" containerName="extract-content" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.100262 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" containerName="extract-content" Dec 10 19:21:13 crc kubenswrapper[4828]: E1210 19:21:13.100283 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd2aa65d-1eae-421a-9198-684337484b92" containerName="init" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.100289 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd2aa65d-1eae-421a-9198-684337484b92" containerName="init" Dec 10 19:21:13 crc kubenswrapper[4828]: E1210 19:21:13.100305 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" containerName="registry-server" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.100311 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" containerName="registry-server" Dec 10 19:21:13 crc kubenswrapper[4828]: E1210 19:21:13.100327 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5296458-e976-4d87-a73f-3fb10cddd3d0" containerName="glance-httpd" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.100335 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5296458-e976-4d87-a73f-3fb10cddd3d0" containerName="glance-httpd" Dec 10 19:21:13 crc kubenswrapper[4828]: E1210 19:21:13.100346 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53369a6e-a6da-4d63-aca6-e1bec8ab047e" containerName="glance-log" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.100352 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="53369a6e-a6da-4d63-aca6-e1bec8ab047e" containerName="glance-log" Dec 10 19:21:13 crc kubenswrapper[4828]: E1210 19:21:13.100368 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" containerName="extract-utilities" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.100375 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" containerName="extract-utilities" Dec 10 19:21:13 crc kubenswrapper[4828]: E1210 19:21:13.100389 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd2aa65d-1eae-421a-9198-684337484b92" containerName="dnsmasq-dns" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.100395 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd2aa65d-1eae-421a-9198-684337484b92" containerName="dnsmasq-dns" Dec 10 19:21:13 crc kubenswrapper[4828]: E1210 19:21:13.100419 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53369a6e-a6da-4d63-aca6-e1bec8ab047e" containerName="glance-httpd" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.100427 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="53369a6e-a6da-4d63-aca6-e1bec8ab047e" containerName="glance-httpd" Dec 10 19:21:13 crc kubenswrapper[4828]: E1210 19:21:13.100440 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5296458-e976-4d87-a73f-3fb10cddd3d0" containerName="glance-log" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.100446 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5296458-e976-4d87-a73f-3fb10cddd3d0" containerName="glance-log" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.100670 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd2aa65d-1eae-421a-9198-684337484b92" containerName="dnsmasq-dns" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.100692 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="53369a6e-a6da-4d63-aca6-e1bec8ab047e" containerName="glance-httpd" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.100709 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5296458-e976-4d87-a73f-3fb10cddd3d0" containerName="glance-log" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.100718 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5296458-e976-4d87-a73f-3fb10cddd3d0" containerName="glance-httpd" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.100728 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="53369a6e-a6da-4d63-aca6-e1bec8ab047e" containerName="glance-log" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.100734 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" containerName="registry-server" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.106376 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.111105 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.111729 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.144211 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.144958 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.214487 4828 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.232615 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53369a6e-a6da-4d63-aca6-e1bec8ab047e" (UID: "53369a6e-a6da-4d63-aca6-e1bec8ab047e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.251678 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-scripts\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.251979 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "53369a6e-a6da-4d63-aca6-e1bec8ab047e" (UID: "53369a6e-a6da-4d63-aca6-e1bec8ab047e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.252021 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.252220 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-config-data\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.252295 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.252361 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89ssd\" (UniqueName: \"kubernetes.io/projected/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-kube-api-access-89ssd\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.252490 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.252561 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.252637 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-logs\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.252920 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.252936 4828 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.252945 4828 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/53369a6e-a6da-4d63-aca6-e1bec8ab047e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.356865 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.356949 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.357019 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-logs\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.357215 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-scripts\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.357334 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.357481 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-config-data\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.357517 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.357565 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89ssd\" (UniqueName: \"kubernetes.io/projected/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-kube-api-access-89ssd\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.360350 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.362702 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-logs\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.362968 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.364594 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-scripts\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.368683 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.368929 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.370033 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-jhdsp" event={"ID":"df859d5f-d811-4be3-82d8-8327c7cfb851","Type":"ContainerStarted","Data":"33d6df1cc9c23399a3e761f34b73e1aa89ae953a44e490b0c961b3bd25621b7f"} Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.370068 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-jhdsp" event={"ID":"df859d5f-d811-4be3-82d8-8327c7cfb851","Type":"ContainerStarted","Data":"30889fb9e1ba260983cff71222ef34f1fda1fe4d6b4bd6fff98172fe22cade1c"} Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.377576 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-config-data\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.377985 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89ssd\" (UniqueName: \"kubernetes.io/projected/2fffc8e4-c48c-4f89-bc5d-143c3d96a557-kube-api-access-89ssd\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.385713 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"28cb9eb3-e89d-460c-9110-32e241d2a44d","Type":"ContainerStarted","Data":"124b272339a408d11a77542e6626ecc9f6e38933712a145f5cdec62b3e38a870"} Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.404647 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-77bd4457c-f5tqz" event={"ID":"6e68a634-d019-464c-bcf7-c1669cf88fcf","Type":"ContainerStarted","Data":"ad8e39cb38d1ee539fc49d8c3b02108dbfed4889dbb37685c8d3dff78d79fd92"} Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.405144 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-77bd4457c-f5tqz" podUID="6e68a634-d019-464c-bcf7-c1669cf88fcf" containerName="heat-api" containerID="cri-o://ad8e39cb38d1ee539fc49d8c3b02108dbfed4889dbb37685c8d3dff78d79fd92" gracePeriod=60 Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.405478 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-77bd4457c-f5tqz" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.418355 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-jhdsp" podStartSLOduration=9.418336233 podStartE2EDuration="9.418336233s" podCreationTimestamp="2025-12-10 19:21:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:21:13.393565656 +0000 UTC m=+1553.904176661" watchObservedRunningTime="2025-12-10 19:21:13.418336233 +0000 UTC m=+1553.928947228" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.426059 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=4.150825369 podStartE2EDuration="26.426040527s" podCreationTimestamp="2025-12-10 19:20:47 +0000 UTC" firstStartedPulling="2025-12-10 19:20:48.755712033 +0000 UTC m=+1529.266323028" lastFinishedPulling="2025-12-10 19:21:11.030927181 +0000 UTC m=+1551.541538186" observedRunningTime="2025-12-10 19:21:13.412420526 +0000 UTC m=+1553.923031531" watchObservedRunningTime="2025-12-10 19:21:13.426040527 +0000 UTC m=+1553.936651532" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.433469 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"2fffc8e4-c48c-4f89-bc5d-143c3d96a557\") " pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.435121 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" event={"ID":"51c89612-b937-4376-88a8-6623b3ec8d05","Type":"ContainerStarted","Data":"ddb5fd1d65b1ba18b944b5c1ea6a52027d0aaab5345ee825ab1adcb6894e17ce"} Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.435285 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" podUID="51c89612-b937-4376-88a8-6623b3ec8d05" containerName="heat-cfnapi" containerID="cri-o://ddb5fd1d65b1ba18b944b5c1ea6a52027d0aaab5345ee825ab1adcb6894e17ce" gracePeriod=60 Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.435369 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.448421 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" event={"ID":"dbfc3632-aaef-44db-a52a-7403cae01ad4","Type":"ContainerStarted","Data":"f892d7ec2fdbf36dfbfd13f66531c9a22a79836aa361d4bd79769497ad8a5f72"} Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.448726 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" event={"ID":"dbfc3632-aaef-44db-a52a-7403cae01ad4","Type":"ContainerStarted","Data":"88ccd8b0d00966fb6e4dadab1cc9c578a434c80e53a522a22b1f66ab116dfbe2"} Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.450067 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.451341 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.454488 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37649bc1-48ed-4cc3-9b7b-788469c5a989","Type":"ContainerStarted","Data":"b3d3644115949d315422eb108ab764c35379cde648deadfb6e78b9f250f94a14"} Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.476596 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53369a6e-a6da-4d63-aca6-e1bec8ab047e","Type":"ContainerDied","Data":"22aa1decab2296eb980332fae51e49edc430a14cb1184cbbde9de15eed0dc3e6"} Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.476665 4828 scope.go:117] "RemoveContainer" containerID="23700b3985da02baa67603281a49c24c027d44b3fb0df6c6db157900b864bded" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.477246 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.482387 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-77bd4457c-f5tqz" podStartSLOduration=4.856015288 podStartE2EDuration="18.482367668s" podCreationTimestamp="2025-12-10 19:20:55 +0000 UTC" firstStartedPulling="2025-12-10 19:20:57.40641891 +0000 UTC m=+1537.917029915" lastFinishedPulling="2025-12-10 19:21:11.03277129 +0000 UTC m=+1551.543382295" observedRunningTime="2025-12-10 19:21:13.446698554 +0000 UTC m=+1553.957309579" watchObservedRunningTime="2025-12-10 19:21:13.482367668 +0000 UTC m=+1553.992978673" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.509472 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" podStartSLOduration=5.7113430229999995 podStartE2EDuration="19.509452726s" podCreationTimestamp="2025-12-10 19:20:54 +0000 UTC" firstStartedPulling="2025-12-10 19:20:56.877180162 +0000 UTC m=+1537.387791167" lastFinishedPulling="2025-12-10 19:21:10.675289865 +0000 UTC m=+1551.185900870" observedRunningTime="2025-12-10 19:21:13.481039323 +0000 UTC m=+1553.991650328" watchObservedRunningTime="2025-12-10 19:21:13.509452726 +0000 UTC m=+1554.020063731" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.633545 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" podStartSLOduration=10.633482891 podStartE2EDuration="10.633482891s" podCreationTimestamp="2025-12-10 19:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:21:13.531055008 +0000 UTC m=+1554.041666013" watchObservedRunningTime="2025-12-10 19:21:13.633482891 +0000 UTC m=+1554.144093896" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.670648 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-e96e-account-create-update-8kssp"] Dec 10 19:21:13 crc kubenswrapper[4828]: W1210 19:21:13.699620 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f5c519b_682a_430f_b1ed_ee93f6b7c9f2.slice/crio-f760874371f7af577fc8ba1e27ac3d2fc531caebacb84f1ee9140ad0f58df80d WatchSource:0}: Error finding container f760874371f7af577fc8ba1e27ac3d2fc531caebacb84f1ee9140ad0f58df80d: Status 404 returned error can't find the container with id f760874371f7af577fc8ba1e27ac3d2fc531caebacb84f1ee9140ad0f58df80d Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.714226 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-546dfdfc9f-s9ctn"] Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.753772 4828 scope.go:117] "RemoveContainer" containerID="af613c0a15c1188568299015ab706de182e2534b39204ffc56468699e1878789" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.775947 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-28cs6"] Dec 10 19:21:13 crc kubenswrapper[4828]: W1210 19:21:13.813948 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63fd1651_0e27_40c0_b5a4_38ab90a957ca.slice/crio-816072689e5886b9ace475f1dd2b93dc9cad209a730993c27633571a5d189cc6 WatchSource:0}: Error finding container 816072689e5886b9ace475f1dd2b93dc9cad209a730993c27633571a5d189cc6: Status 404 returned error can't find the container with id 816072689e5886b9ace475f1dd2b93dc9cad209a730993c27633571a5d189cc6 Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.837157 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5296458-e976-4d87-a73f-3fb10cddd3d0" path="/var/lib/kubelet/pods/c5296458-e976-4d87-a73f-3fb10cddd3d0/volumes" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.838053 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd2aa65d-1eae-421a-9198-684337484b92" path="/var/lib/kubelet/pods/fd2aa65d-1eae-421a-9198-684337484b92/volumes" Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.846342 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.935962 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:21:13 crc kubenswrapper[4828]: I1210 19:21:13.997998 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-566d76fd44-v6jl7"] Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.032231 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-edd8-account-create-update-f87t8"] Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.047022 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-d78dbcd66-dq2t9"] Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.148743 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-59c4ffddfc-nf4zg"] Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.164851 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-c67d-account-create-update-d6zs6"] Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.195640 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5tqrl"] Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.214443 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.217083 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.224341 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.232857 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.258684 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.389021 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.428292 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c08f474-3395-463f-ac2f-559aa97bb021-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.428352 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c08f474-3395-463f-ac2f-559aa97bb021-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.428398 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3c08f474-3395-463f-ac2f-559aa97bb021-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.428501 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c08f474-3395-463f-ac2f-559aa97bb021-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.428544 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.428571 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-757qc\" (UniqueName: \"kubernetes.io/projected/3c08f474-3395-463f-ac2f-559aa97bb021-kube-api-access-757qc\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.428599 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c08f474-3395-463f-ac2f-559aa97bb021-logs\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.428698 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c08f474-3395-463f-ac2f-559aa97bb021-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.500128 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c67d-account-create-update-d6zs6" event={"ID":"47be5ead-d925-4058-a390-88426775ebfa","Type":"ContainerStarted","Data":"ff4b72097f5fa980a04470e08de4f26a6d0b0ce77a977debbf8fa4729d8b6ef5"} Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.502711 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e96e-account-create-update-8kssp" event={"ID":"1f5c519b-682a-430f-b1ed-ee93f6b7c9f2","Type":"ContainerStarted","Data":"2c3945b533871f06f437af5630054bead64d7ada9fe59b3de65d4ba734c03760"} Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.502743 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e96e-account-create-update-8kssp" event={"ID":"1f5c519b-682a-430f-b1ed-ee93f6b7c9f2","Type":"ContainerStarted","Data":"f760874371f7af577fc8ba1e27ac3d2fc531caebacb84f1ee9140ad0f58df80d"} Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.505360 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" event={"ID":"296faf68-6251-48d1-bc36-026a31c00393","Type":"ContainerStarted","Data":"0bb5e90efc664c51989f69ee2d7abf7b0061f2d85cb18eaf58cfdd7c0d1065bf"} Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.507250 4828 generic.go:334] "Generic (PLEG): container finished" podID="8cb3c5d9-61fb-415a-a644-75803d3f7c8f" containerID="98e81aab581da45b5907a2bcaf76c5664a8e065cd952317d5fb3991b83648371" exitCode=137 Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.507321 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8cb3c5d9-61fb-415a-a644-75803d3f7c8f","Type":"ContainerDied","Data":"98e81aab581da45b5907a2bcaf76c5664a8e065cd952317d5fb3991b83648371"} Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.513208 4828 generic.go:334] "Generic (PLEG): container finished" podID="dbfc3632-aaef-44db-a52a-7403cae01ad4" containerID="f892d7ec2fdbf36dfbfd13f66531c9a22a79836aa361d4bd79769497ad8a5f72" exitCode=1 Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.513289 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" event={"ID":"dbfc3632-aaef-44db-a52a-7403cae01ad4","Type":"ContainerDied","Data":"f892d7ec2fdbf36dfbfd13f66531c9a22a79836aa361d4bd79769497ad8a5f72"} Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.514391 4828 scope.go:117] "RemoveContainer" containerID="f892d7ec2fdbf36dfbfd13f66531c9a22a79836aa361d4bd79769497ad8a5f72" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.517939 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5tqrl" event={"ID":"d24aa875-f0bd-4e7d-a985-c633a246fb30","Type":"ContainerStarted","Data":"f8e3cba33baef3ea5e00d70fa7e64264e5324cd3f5d1e1da9a595c0cbfd31f12"} Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.519625 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2fffc8e4-c48c-4f89-bc5d-143c3d96a557","Type":"ContainerStarted","Data":"5093462ad673bf4ac3831a0a2ef24b0c87ccef128f8808e11e3474ee7e9e4053"} Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.523643 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-28cs6" event={"ID":"b544108c-9253-40f1-a5ed-8456173a3e68","Type":"ContainerStarted","Data":"6c8ce06fc3243cd513cd787f8c8c245105f1ff74204738c70ce23c131ae91698"} Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.537527 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-e96e-account-create-update-8kssp" podStartSLOduration=9.537506143 podStartE2EDuration="9.537506143s" podCreationTimestamp="2025-12-10 19:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:21:14.529814319 +0000 UTC m=+1555.040425344" watchObservedRunningTime="2025-12-10 19:21:14.537506143 +0000 UTC m=+1555.048117148" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.538745 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c08f474-3395-463f-ac2f-559aa97bb021-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.539979 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.540020 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-757qc\" (UniqueName: \"kubernetes.io/projected/3c08f474-3395-463f-ac2f-559aa97bb021-kube-api-access-757qc\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.540050 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c08f474-3395-463f-ac2f-559aa97bb021-logs\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.540186 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c08f474-3395-463f-ac2f-559aa97bb021-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.540309 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c08f474-3395-463f-ac2f-559aa97bb021-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.540346 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c08f474-3395-463f-ac2f-559aa97bb021-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.540408 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3c08f474-3395-463f-ac2f-559aa97bb021-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.540563 4828 generic.go:334] "Generic (PLEG): container finished" podID="df859d5f-d811-4be3-82d8-8327c7cfb851" containerID="33d6df1cc9c23399a3e761f34b73e1aa89ae953a44e490b0c961b3bd25621b7f" exitCode=0 Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.540703 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-jhdsp" event={"ID":"df859d5f-d811-4be3-82d8-8327c7cfb851","Type":"ContainerDied","Data":"33d6df1cc9c23399a3e761f34b73e1aa89ae953a44e490b0c961b3bd25621b7f"} Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.541020 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3c08f474-3395-463f-ac2f-559aa97bb021-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.541392 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.541734 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3c08f474-3395-463f-ac2f-559aa97bb021-logs\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.564059 4828 generic.go:334] "Generic (PLEG): container finished" podID="6e68a634-d019-464c-bcf7-c1669cf88fcf" containerID="ad8e39cb38d1ee539fc49d8c3b02108dbfed4889dbb37685c8d3dff78d79fd92" exitCode=0 Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.564257 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-77bd4457c-f5tqz" event={"ID":"6e68a634-d019-464c-bcf7-c1669cf88fcf","Type":"ContainerDied","Data":"ad8e39cb38d1ee539fc49d8c3b02108dbfed4889dbb37685c8d3dff78d79fd92"} Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.564559 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c08f474-3395-463f-ac2f-559aa97bb021-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.564599 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c08f474-3395-463f-ac2f-559aa97bb021-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.565202 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c08f474-3395-463f-ac2f-559aa97bb021-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.574361 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-566d76fd44-v6jl7" event={"ID":"63fd1651-0e27-40c0-b5a4-38ab90a957ca","Type":"ContainerStarted","Data":"816072689e5886b9ace475f1dd2b93dc9cad209a730993c27633571a5d189cc6"} Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.577089 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c08f474-3395-463f-ac2f-559aa97bb021-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.578119 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-d78dbcd66-dq2t9" event={"ID":"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b","Type":"ContainerStarted","Data":"b962bcb63927ba9ce27b85c9abfe3ecabe99b6d3c11f40facbbaeea501982b41"} Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.593085 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-edd8-account-create-update-f87t8" event={"ID":"b19ce251-54b7-40f3-bbe0-6afb6064b307","Type":"ContainerStarted","Data":"cdfddfc4cb8804e6123be5a09adae2421022371cb6d3f95765e08c364825f808"} Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.606818 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-757qc\" (UniqueName: \"kubernetes.io/projected/3c08f474-3395-463f-ac2f-559aa97bb021-kube-api-access-757qc\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.612575 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-59c4ffddfc-nf4zg" event={"ID":"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709","Type":"ContainerStarted","Data":"9449e358d2c50c03042939037ad52b848ccaf0d226e7ddd9f232bb8cc911cb58"} Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.623622 4828 generic.go:334] "Generic (PLEG): container finished" podID="51c89612-b937-4376-88a8-6623b3ec8d05" containerID="ddb5fd1d65b1ba18b944b5c1ea6a52027d0aaab5345ee825ab1adcb6894e17ce" exitCode=0 Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.624231 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" event={"ID":"51c89612-b937-4376-88a8-6623b3ec8d05","Type":"ContainerDied","Data":"ddb5fd1d65b1ba18b944b5c1ea6a52027d0aaab5345ee825ab1adcb6894e17ce"} Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.666586 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37649bc1-48ed-4cc3-9b7b-788469c5a989","Type":"ContainerStarted","Data":"1dfa7423ff1d5a745b35b1ffa2137f0fe17c49f0bfaec5617f8817a60827ac18"} Dec 10 19:21:14 crc kubenswrapper[4828]: I1210 19:21:14.690615 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"3c08f474-3395-463f-ac2f-559aa97bb021\") " pod="openstack/glance-default-internal-api-0" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.068732 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.223650 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.230298 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-77bd4457c-f5tqz" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.241469 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.261118 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-etc-machine-id\") pod \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.261372 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-config-data\") pod \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.261403 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptvb6\" (UniqueName: \"kubernetes.io/projected/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-kube-api-access-ptvb6\") pod \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.261460 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-logs\") pod \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.261489 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-scripts\") pod \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.261560 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-combined-ca-bundle\") pod \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.261588 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-config-data-custom\") pod \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\" (UID: \"8cb3c5d9-61fb-415a-a644-75803d3f7c8f\") " Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.264534 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8cb3c5d9-61fb-415a-a644-75803d3f7c8f" (UID: "8cb3c5d9-61fb-415a-a644-75803d3f7c8f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.273995 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-logs" (OuterVolumeSpecName: "logs") pod "8cb3c5d9-61fb-415a-a644-75803d3f7c8f" (UID: "8cb3c5d9-61fb-415a-a644-75803d3f7c8f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.340716 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8cb3c5d9-61fb-415a-a644-75803d3f7c8f" (UID: "8cb3c5d9-61fb-415a-a644-75803d3f7c8f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.352653 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-kube-api-access-ptvb6" (OuterVolumeSpecName: "kube-api-access-ptvb6") pod "8cb3c5d9-61fb-415a-a644-75803d3f7c8f" (UID: "8cb3c5d9-61fb-415a-a644-75803d3f7c8f"). InnerVolumeSpecName "kube-api-access-ptvb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.354212 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-scripts" (OuterVolumeSpecName: "scripts") pod "8cb3c5d9-61fb-415a-a644-75803d3f7c8f" (UID: "8cb3c5d9-61fb-415a-a644-75803d3f7c8f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.363692 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-config-data-custom\") pod \"51c89612-b937-4376-88a8-6623b3ec8d05\" (UID: \"51c89612-b937-4376-88a8-6623b3ec8d05\") " Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.363846 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jmgb\" (UniqueName: \"kubernetes.io/projected/51c89612-b937-4376-88a8-6623b3ec8d05-kube-api-access-5jmgb\") pod \"51c89612-b937-4376-88a8-6623b3ec8d05\" (UID: \"51c89612-b937-4376-88a8-6623b3ec8d05\") " Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.363905 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-combined-ca-bundle\") pod \"51c89612-b937-4376-88a8-6623b3ec8d05\" (UID: \"51c89612-b937-4376-88a8-6623b3ec8d05\") " Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.363965 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-config-data-custom\") pod \"6e68a634-d019-464c-bcf7-c1669cf88fcf\" (UID: \"6e68a634-d019-464c-bcf7-c1669cf88fcf\") " Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.364033 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-config-data\") pod \"51c89612-b937-4376-88a8-6623b3ec8d05\" (UID: \"51c89612-b937-4376-88a8-6623b3ec8d05\") " Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.364088 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-combined-ca-bundle\") pod \"6e68a634-d019-464c-bcf7-c1669cf88fcf\" (UID: \"6e68a634-d019-464c-bcf7-c1669cf88fcf\") " Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.364118 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgr5d\" (UniqueName: \"kubernetes.io/projected/6e68a634-d019-464c-bcf7-c1669cf88fcf-kube-api-access-fgr5d\") pod \"6e68a634-d019-464c-bcf7-c1669cf88fcf\" (UID: \"6e68a634-d019-464c-bcf7-c1669cf88fcf\") " Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.364147 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-config-data\") pod \"6e68a634-d019-464c-bcf7-c1669cf88fcf\" (UID: \"6e68a634-d019-464c-bcf7-c1669cf88fcf\") " Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.368389 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptvb6\" (UniqueName: \"kubernetes.io/projected/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-kube-api-access-ptvb6\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.368444 4828 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.368461 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.368471 4828 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.368937 4828 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.390130 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "51c89612-b937-4376-88a8-6623b3ec8d05" (UID: "51c89612-b937-4376-88a8-6623b3ec8d05"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.390287 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e68a634-d019-464c-bcf7-c1669cf88fcf-kube-api-access-fgr5d" (OuterVolumeSpecName: "kube-api-access-fgr5d") pod "6e68a634-d019-464c-bcf7-c1669cf88fcf" (UID: "6e68a634-d019-464c-bcf7-c1669cf88fcf"). InnerVolumeSpecName "kube-api-access-fgr5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.390566 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6e68a634-d019-464c-bcf7-c1669cf88fcf" (UID: "6e68a634-d019-464c-bcf7-c1669cf88fcf"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.413380 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51c89612-b937-4376-88a8-6623b3ec8d05-kube-api-access-5jmgb" (OuterVolumeSpecName: "kube-api-access-5jmgb") pod "51c89612-b937-4376-88a8-6623b3ec8d05" (UID: "51c89612-b937-4376-88a8-6623b3ec8d05"). InnerVolumeSpecName "kube-api-access-5jmgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.470703 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgr5d\" (UniqueName: \"kubernetes.io/projected/6e68a634-d019-464c-bcf7-c1669cf88fcf-kube-api-access-fgr5d\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.471030 4828 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.471043 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jmgb\" (UniqueName: \"kubernetes.io/projected/51c89612-b937-4376-88a8-6623b3ec8d05-kube-api-access-5jmgb\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.471057 4828 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.564111 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.689092 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-77bd4457c-f5tqz" event={"ID":"6e68a634-d019-464c-bcf7-c1669cf88fcf","Type":"ContainerDied","Data":"77639af20d9aeb386bf2fa122b3d2ec2c301101e6cf5101fd48508a1e778363a"} Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.689181 4828 scope.go:117] "RemoveContainer" containerID="ad8e39cb38d1ee539fc49d8c3b02108dbfed4889dbb37685c8d3dff78d79fd92" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.689516 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-77bd4457c-f5tqz" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.695686 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c67d-account-create-update-d6zs6" event={"ID":"47be5ead-d925-4058-a390-88426775ebfa","Type":"ContainerStarted","Data":"1b6601f4598967d3b11414c3a4ffa6142b58840c11402f805f095360b0f6cc2e"} Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.710574 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8cb3c5d9-61fb-415a-a644-75803d3f7c8f","Type":"ContainerDied","Data":"dc140dccad70f788c8579c38684802e86b64741f2ccc404dfd77524bfc74d057"} Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.710654 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.717480 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-c67d-account-create-update-d6zs6" podStartSLOduration=10.717464544 podStartE2EDuration="10.717464544s" podCreationTimestamp="2025-12-10 19:21:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:21:15.717059593 +0000 UTC m=+1556.227670598" watchObservedRunningTime="2025-12-10 19:21:15.717464544 +0000 UTC m=+1556.228075549" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.726698 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" event={"ID":"51c89612-b937-4376-88a8-6623b3ec8d05","Type":"ContainerDied","Data":"2cead302fa78ebee00121c74029cb5f302dc7a3ce5f2c3981c0fd7adbc9b2634"} Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.726783 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6f7b6d44ff-rhjwx" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.731907 4828 generic.go:334] "Generic (PLEG): container finished" podID="1f5c519b-682a-430f-b1ed-ee93f6b7c9f2" containerID="2c3945b533871f06f437af5630054bead64d7ada9fe59b3de65d4ba734c03760" exitCode=0 Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.731996 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e96e-account-create-update-8kssp" event={"ID":"1f5c519b-682a-430f-b1ed-ee93f6b7c9f2","Type":"ContainerDied","Data":"2c3945b533871f06f437af5630054bead64d7ada9fe59b3de65d4ba734c03760"} Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.814006 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53369a6e-a6da-4d63-aca6-e1bec8ab047e" path="/var/lib/kubelet/pods/53369a6e-a6da-4d63-aca6-e1bec8ab047e/volumes" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.823077 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8cb3c5d9-61fb-415a-a644-75803d3f7c8f" (UID: "8cb3c5d9-61fb-415a-a644-75803d3f7c8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:15 crc kubenswrapper[4828]: I1210 19:21:15.891998 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:16 crc kubenswrapper[4828]: I1210 19:21:16.332674 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "51c89612-b937-4376-88a8-6623b3ec8d05" (UID: "51c89612-b937-4376-88a8-6623b3ec8d05"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:16 crc kubenswrapper[4828]: I1210 19:21:16.337663 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:16 crc kubenswrapper[4828]: I1210 19:21:16.500092 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-config-data" (OuterVolumeSpecName: "config-data") pod "8cb3c5d9-61fb-415a-a644-75803d3f7c8f" (UID: "8cb3c5d9-61fb-415a-a644-75803d3f7c8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:16 crc kubenswrapper[4828]: I1210 19:21:16.518138 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-config-data" (OuterVolumeSpecName: "config-data") pod "6e68a634-d019-464c-bcf7-c1669cf88fcf" (UID: "6e68a634-d019-464c-bcf7-c1669cf88fcf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:16 crc kubenswrapper[4828]: I1210 19:21:16.546980 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:16 crc kubenswrapper[4828]: I1210 19:21:16.547303 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cb3c5d9-61fb-415a-a644-75803d3f7c8f-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:16 crc kubenswrapper[4828]: I1210 19:21:16.573028 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e68a634-d019-464c-bcf7-c1669cf88fcf" (UID: "6e68a634-d019-464c-bcf7-c1669cf88fcf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:16 crc kubenswrapper[4828]: I1210 19:21:16.615988 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-config-data" (OuterVolumeSpecName: "config-data") pod "51c89612-b937-4376-88a8-6623b3ec8d05" (UID: "51c89612-b937-4376-88a8-6623b3ec8d05"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:16 crc kubenswrapper[4828]: I1210 19:21:16.651760 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51c89612-b937-4376-88a8-6623b3ec8d05-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:16 crc kubenswrapper[4828]: I1210 19:21:16.651991 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e68a634-d019-464c-bcf7-c1669cf88fcf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:16 crc kubenswrapper[4828]: I1210 19:21:16.865693 4828 generic.go:334] "Generic (PLEG): container finished" podID="b544108c-9253-40f1-a5ed-8456173a3e68" containerID="e6361c7b85657f7de6ac78e64a0b6a2bd350a593ed7056221d07e6c4b4c70376" exitCode=0 Dec 10 19:21:16 crc kubenswrapper[4828]: I1210 19:21:16.912672 4828 generic.go:334] "Generic (PLEG): container finished" podID="d24aa875-f0bd-4e7d-a985-c633a246fb30" containerID="cca446d297b8461f3e8fd45046e8da8e6e5e9193cadc282562fe2e510f32cc05" exitCode=0 Dec 10 19:21:16 crc kubenswrapper[4828]: I1210 19:21:16.928613 4828 generic.go:334] "Generic (PLEG): container finished" podID="47be5ead-d925-4058-a390-88426775ebfa" containerID="1b6601f4598967d3b11414c3a4ffa6142b58840c11402f805f095360b0f6cc2e" exitCode=0 Dec 10 19:21:16 crc kubenswrapper[4828]: I1210 19:21:16.936162 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-59c4ffddfc-nf4zg" podStartSLOduration=10.936146899 podStartE2EDuration="10.936146899s" podCreationTimestamp="2025-12-10 19:21:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:21:16.88216794 +0000 UTC m=+1557.392778935" watchObservedRunningTime="2025-12-10 19:21:16.936146899 +0000 UTC m=+1557.446757904" Dec 10 19:21:16 crc kubenswrapper[4828]: I1210 19:21:16.948032 4828 generic.go:334] "Generic (PLEG): container finished" podID="b19ce251-54b7-40f3-bbe0-6afb6064b307" containerID="bc65d15e3a6a8d0cd222585dce602e371e206cbf5ac9b84fb263d5b0433ef7d7" exitCode=0 Dec 10 19:21:16 crc kubenswrapper[4828]: I1210 19:21:16.952036 4828 generic.go:334] "Generic (PLEG): container finished" podID="dbfc3632-aaef-44db-a52a-7403cae01ad4" containerID="7761a7fab7f1757991c79aa7082709c281bce53ed2c6d4d94fb4ef159ff329b9" exitCode=1 Dec 10 19:21:16 crc kubenswrapper[4828]: I1210 19:21:16.952744 4828 scope.go:117] "RemoveContainer" containerID="7761a7fab7f1757991c79aa7082709c281bce53ed2c6d4d94fb4ef159ff329b9" Dec 10 19:21:16 crc kubenswrapper[4828]: E1210 19:21:16.953077 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-6dfc7956d8-nq8qr_openstack(dbfc3632-aaef-44db-a52a-7403cae01ad4)\"" pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" podUID="dbfc3632-aaef-44db-a52a-7403cae01ad4" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.019997 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" podStartSLOduration=11.019975609 podStartE2EDuration="11.019975609s" podCreationTimestamp="2025-12-10 19:21:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:21:17.017195105 +0000 UTC m=+1557.527806120" watchObservedRunningTime="2025-12-10 19:21:17.019975609 +0000 UTC m=+1557.530586614" Dec 10 19:21:17 crc kubenswrapper[4828]: E1210 19:21:17.128115 4828 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.335s" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.128234 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-59c4ffddfc-nf4zg" event={"ID":"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709","Type":"ContainerStarted","Data":"4e62436d5f048b277525938d4b6e74b8b4d6af7192d374e82a9fb90de498c848"} Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.128401 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.128420 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" event={"ID":"296faf68-6251-48d1-bc36-026a31c00393","Type":"ContainerStarted","Data":"b514a4dbaafec84cc689eb9deaf63a5685ff0cb779acc7f815431dae3f169013"} Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.128604 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.128616 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.128624 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.128634 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.128648 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-28cs6" event={"ID":"b544108c-9253-40f1-a5ed-8456173a3e68","Type":"ContainerDied","Data":"e6361c7b85657f7de6ac78e64a0b6a2bd350a593ed7056221d07e6c4b4c70376"} Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.128666 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-jhdsp" event={"ID":"df859d5f-d811-4be3-82d8-8327c7cfb851","Type":"ContainerDied","Data":"30889fb9e1ba260983cff71222ef34f1fda1fe4d6b4bd6fff98172fe22cade1c"} Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.128679 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30889fb9e1ba260983cff71222ef34f1fda1fe4d6b4bd6fff98172fe22cade1c" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.128690 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-566d76fd44-v6jl7" event={"ID":"63fd1651-0e27-40c0-b5a4-38ab90a957ca","Type":"ContainerStarted","Data":"e36067d92c45d3ac2f0eaae0855f770be2bbe334104e3cbe368c63096553a519"} Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.128698 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5tqrl" event={"ID":"d24aa875-f0bd-4e7d-a985-c633a246fb30","Type":"ContainerDied","Data":"cca446d297b8461f3e8fd45046e8da8e6e5e9193cadc282562fe2e510f32cc05"} Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.129170 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-d78dbcd66-dq2t9" event={"ID":"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b","Type":"ContainerStarted","Data":"3f7f00335c86a1e5b99e0e3677a9c6cf168454f042fc2f5a720193b77fff2c7e"} Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.129183 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c67d-account-create-update-d6zs6" event={"ID":"47be5ead-d925-4058-a390-88426775ebfa","Type":"ContainerDied","Data":"1b6601f4598967d3b11414c3a4ffa6142b58840c11402f805f095360b0f6cc2e"} Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.129196 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-edd8-account-create-update-f87t8" event={"ID":"b19ce251-54b7-40f3-bbe0-6afb6064b307","Type":"ContainerDied","Data":"bc65d15e3a6a8d0cd222585dce602e371e206cbf5ac9b84fb263d5b0433ef7d7"} Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.129207 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" event={"ID":"dbfc3632-aaef-44db-a52a-7403cae01ad4","Type":"ContainerDied","Data":"7761a7fab7f1757991c79aa7082709c281bce53ed2c6d4d94fb4ef159ff329b9"} Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.129222 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37649bc1-48ed-4cc3-9b7b-788469c5a989","Type":"ContainerStarted","Data":"b78a3e2e45e64cfdc98c611229d668709a2ba675192f598624e1f93a72b18fb9"} Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.129232 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2fffc8e4-c48c-4f89-bc5d-143c3d96a557","Type":"ContainerStarted","Data":"46885bc928f972c3fbf34cb5470187b28b2910cd4f21b2a57a1cac4fa8a3520d"} Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.129241 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3c08f474-3395-463f-ac2f-559aa97bb021","Type":"ContainerStarted","Data":"aeeb87de17f9e55db02b32045297b691a254204618818b2585a7e7935ca0693d"} Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.187785 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-566d76fd44-v6jl7" podStartSLOduration=14.187765653 podStartE2EDuration="14.187765653s" podCreationTimestamp="2025-12-10 19:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:21:17.165993936 +0000 UTC m=+1557.676604931" watchObservedRunningTime="2025-12-10 19:21:17.187765653 +0000 UTC m=+1557.698376658" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.193981 4828 scope.go:117] "RemoveContainer" containerID="98e81aab581da45b5907a2bcaf76c5664a8e065cd952317d5fb3991b83648371" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.210383 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-d78dbcd66-dq2t9" podStartSLOduration=14.210364991 podStartE2EDuration="14.210364991s" podCreationTimestamp="2025-12-10 19:21:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:21:17.194690346 +0000 UTC m=+1557.705301351" watchObservedRunningTime="2025-12-10 19:21:17.210364991 +0000 UTC m=+1557.720975996" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.232474 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-jhdsp" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.352935 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.376190 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df859d5f-d811-4be3-82d8-8327c7cfb851-operator-scripts\") pod \"df859d5f-d811-4be3-82d8-8327c7cfb851\" (UID: \"df859d5f-d811-4be3-82d8-8327c7cfb851\") " Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.376554 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lljp\" (UniqueName: \"kubernetes.io/projected/df859d5f-d811-4be3-82d8-8327c7cfb851-kube-api-access-7lljp\") pod \"df859d5f-d811-4be3-82d8-8327c7cfb851\" (UID: \"df859d5f-d811-4be3-82d8-8327c7cfb851\") " Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.377500 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df859d5f-d811-4be3-82d8-8327c7cfb851-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "df859d5f-d811-4be3-82d8-8327c7cfb851" (UID: "df859d5f-d811-4be3-82d8-8327c7cfb851"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.385511 4828 scope.go:117] "RemoveContainer" containerID="00139c641c2eb6e20a80832cd0c631ec70582d91abe9adaad0315426b3e6248a" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.398874 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.421546 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:21:17 crc kubenswrapper[4828]: E1210 19:21:17.422363 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e68a634-d019-464c-bcf7-c1669cf88fcf" containerName="heat-api" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.422381 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e68a634-d019-464c-bcf7-c1669cf88fcf" containerName="heat-api" Dec 10 19:21:17 crc kubenswrapper[4828]: E1210 19:21:17.422439 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df859d5f-d811-4be3-82d8-8327c7cfb851" containerName="mariadb-database-create" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.422447 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="df859d5f-d811-4be3-82d8-8327c7cfb851" containerName="mariadb-database-create" Dec 10 19:21:17 crc kubenswrapper[4828]: E1210 19:21:17.422465 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51c89612-b937-4376-88a8-6623b3ec8d05" containerName="heat-cfnapi" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.422473 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="51c89612-b937-4376-88a8-6623b3ec8d05" containerName="heat-cfnapi" Dec 10 19:21:17 crc kubenswrapper[4828]: E1210 19:21:17.422485 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cb3c5d9-61fb-415a-a644-75803d3f7c8f" containerName="cinder-api" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.422493 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cb3c5d9-61fb-415a-a644-75803d3f7c8f" containerName="cinder-api" Dec 10 19:21:17 crc kubenswrapper[4828]: E1210 19:21:17.422507 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cb3c5d9-61fb-415a-a644-75803d3f7c8f" containerName="cinder-api-log" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.422514 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cb3c5d9-61fb-415a-a644-75803d3f7c8f" containerName="cinder-api-log" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.422791 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="df859d5f-d811-4be3-82d8-8327c7cfb851" containerName="mariadb-database-create" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.422828 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="51c89612-b937-4376-88a8-6623b3ec8d05" containerName="heat-cfnapi" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.422840 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cb3c5d9-61fb-415a-a644-75803d3f7c8f" containerName="cinder-api" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.422870 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cb3c5d9-61fb-415a-a644-75803d3f7c8f" containerName="cinder-api-log" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.422879 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e68a634-d019-464c-bcf7-c1669cf88fcf" containerName="heat-api" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.424418 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.442400 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df859d5f-d811-4be3-82d8-8327c7cfb851-kube-api-access-7lljp" (OuterVolumeSpecName: "kube-api-access-7lljp") pod "df859d5f-d811-4be3-82d8-8327c7cfb851" (UID: "df859d5f-d811-4be3-82d8-8327c7cfb851"). InnerVolumeSpecName "kube-api-access-7lljp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.449272 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.449556 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.449744 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.464343 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-77bd4457c-f5tqz"] Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.485089 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/71db2c7f-94ac-455b-9142-c63bb53ed942-etc-machine-id\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.485403 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-config-data\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.485424 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.485491 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71db2c7f-94ac-455b-9142-c63bb53ed942-logs\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.485514 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.485530 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-247bj\" (UniqueName: \"kubernetes.io/projected/71db2c7f-94ac-455b-9142-c63bb53ed942-kube-api-access-247bj\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.485707 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-public-tls-certs\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.485747 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-config-data-custom\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.485783 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-scripts\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.485879 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lljp\" (UniqueName: \"kubernetes.io/projected/df859d5f-d811-4be3-82d8-8327c7cfb851-kube-api-access-7lljp\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.485891 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/df859d5f-d811-4be3-82d8-8327c7cfb851-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.490034 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-77bd4457c-f5tqz"] Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.553444 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.581774 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6f7b6d44ff-rhjwx"] Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.584982 4828 scope.go:117] "RemoveContainer" containerID="ddb5fd1d65b1ba18b944b5c1ea6a52027d0aaab5345ee825ab1adcb6894e17ce" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.588887 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-config-data\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.588924 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.589042 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71db2c7f-94ac-455b-9142-c63bb53ed942-logs\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.589082 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.589101 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-247bj\" (UniqueName: \"kubernetes.io/projected/71db2c7f-94ac-455b-9142-c63bb53ed942-kube-api-access-247bj\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.589216 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-public-tls-certs\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.589252 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-config-data-custom\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.589306 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-scripts\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.589393 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/71db2c7f-94ac-455b-9142-c63bb53ed942-etc-machine-id\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.589542 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/71db2c7f-94ac-455b-9142-c63bb53ed942-etc-machine-id\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.594693 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71db2c7f-94ac-455b-9142-c63bb53ed942-logs\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.609387 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-scripts\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.609425 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-public-tls-certs\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.633352 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.641277 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.641999 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-config-data\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.652181 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-247bj\" (UniqueName: \"kubernetes.io/projected/71db2c7f-94ac-455b-9142-c63bb53ed942-kube-api-access-247bj\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.659336 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-6f7b6d44ff-rhjwx"] Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.670718 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/71db2c7f-94ac-455b-9142-c63bb53ed942-config-data-custom\") pod \"cinder-api-0\" (UID: \"71db2c7f-94ac-455b-9142-c63bb53ed942\") " pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.749012 4828 scope.go:117] "RemoveContainer" containerID="f892d7ec2fdbf36dfbfd13f66531c9a22a79836aa361d4bd79769497ad8a5f72" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.812524 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.815290 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51c89612-b937-4376-88a8-6623b3ec8d05" path="/var/lib/kubelet/pods/51c89612-b937-4376-88a8-6623b3ec8d05/volumes" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.816201 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e68a634-d019-464c-bcf7-c1669cf88fcf" path="/var/lib/kubelet/pods/6e68a634-d019-464c-bcf7-c1669cf88fcf/volumes" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.816874 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cb3c5d9-61fb-415a-a644-75803d3f7c8f" path="/var/lib/kubelet/pods/8cb3c5d9-61fb-415a-a644-75803d3f7c8f/volumes" Dec 10 19:21:17 crc kubenswrapper[4828]: I1210 19:21:17.994982 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e96e-account-create-update-8kssp" Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.108383 4828 scope.go:117] "RemoveContainer" containerID="7761a7fab7f1757991c79aa7082709c281bce53ed2c6d4d94fb4ef159ff329b9" Dec 10 19:21:18 crc kubenswrapper[4828]: E1210 19:21:18.108879 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-6dfc7956d8-nq8qr_openstack(dbfc3632-aaef-44db-a52a-7403cae01ad4)\"" pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" podUID="dbfc3632-aaef-44db-a52a-7403cae01ad4" Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.110222 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f5c519b-682a-430f-b1ed-ee93f6b7c9f2-operator-scripts\") pod \"1f5c519b-682a-430f-b1ed-ee93f6b7c9f2\" (UID: \"1f5c519b-682a-430f-b1ed-ee93f6b7c9f2\") " Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.112155 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zjkp\" (UniqueName: \"kubernetes.io/projected/1f5c519b-682a-430f-b1ed-ee93f6b7c9f2-kube-api-access-2zjkp\") pod \"1f5c519b-682a-430f-b1ed-ee93f6b7c9f2\" (UID: \"1f5c519b-682a-430f-b1ed-ee93f6b7c9f2\") " Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.110759 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f5c519b-682a-430f-b1ed-ee93f6b7c9f2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1f5c519b-682a-430f-b1ed-ee93f6b7c9f2" (UID: "1f5c519b-682a-430f-b1ed-ee93f6b7c9f2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.113109 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f5c519b-682a-430f-b1ed-ee93f6b7c9f2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.124493 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e96e-account-create-update-8kssp" Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.124494 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e96e-account-create-update-8kssp" event={"ID":"1f5c519b-682a-430f-b1ed-ee93f6b7c9f2","Type":"ContainerDied","Data":"f760874371f7af577fc8ba1e27ac3d2fc531caebacb84f1ee9140ad0f58df80d"} Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.125623 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f760874371f7af577fc8ba1e27ac3d2fc531caebacb84f1ee9140ad0f58df80d" Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.128893 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37649bc1-48ed-4cc3-9b7b-788469c5a989","Type":"ContainerStarted","Data":"90ce2e9ec5dd43642c4b2efb3407b0bda330694048ed09ab454805247c0c9905"} Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.143626 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3c08f474-3395-463f-ac2f-559aa97bb021","Type":"ContainerStarted","Data":"f0351cdefb2fc787e023cb1e83f89ab5abbe784125f025237bfcb6fca49da6e3"} Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.147170 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f5c519b-682a-430f-b1ed-ee93f6b7c9f2-kube-api-access-2zjkp" (OuterVolumeSpecName: "kube-api-access-2zjkp") pod "1f5c519b-682a-430f-b1ed-ee93f6b7c9f2" (UID: "1f5c519b-682a-430f-b1ed-ee93f6b7c9f2"). InnerVolumeSpecName "kube-api-access-2zjkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.149723 4828 generic.go:334] "Generic (PLEG): container finished" podID="63fd1651-0e27-40c0-b5a4-38ab90a957ca" containerID="e36067d92c45d3ac2f0eaae0855f770be2bbe334104e3cbe368c63096553a519" exitCode=1 Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.150743 4828 scope.go:117] "RemoveContainer" containerID="e36067d92c45d3ac2f0eaae0855f770be2bbe334104e3cbe368c63096553a519" Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.151196 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-566d76fd44-v6jl7" event={"ID":"63fd1651-0e27-40c0-b5a4-38ab90a957ca","Type":"ContainerDied","Data":"e36067d92c45d3ac2f0eaae0855f770be2bbe334104e3cbe368c63096553a519"} Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.152858 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-jhdsp" Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.233054 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zjkp\" (UniqueName: \"kubernetes.io/projected/1f5c519b-682a-430f-b1ed-ee93f6b7c9f2-kube-api-access-2zjkp\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:18 crc kubenswrapper[4828]: E1210 19:21:18.339573 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf859d5f_d811_4be3_82d8_8327c7cfb851.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf859d5f_d811_4be3_82d8_8327c7cfb851.slice/crio-30889fb9e1ba260983cff71222ef34f1fda1fe4d6b4bd6fff98172fe22cade1c\": RecentStats: unable to find data in memory cache]" Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.422010 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.709931 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.761496 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.761765 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:18 crc kubenswrapper[4828]: I1210 19:21:18.927929 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c67d-account-create-update-d6zs6" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.096142 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47be5ead-d925-4058-a390-88426775ebfa-operator-scripts\") pod \"47be5ead-d925-4058-a390-88426775ebfa\" (UID: \"47be5ead-d925-4058-a390-88426775ebfa\") " Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.096501 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh6vf\" (UniqueName: \"kubernetes.io/projected/47be5ead-d925-4058-a390-88426775ebfa-kube-api-access-lh6vf\") pod \"47be5ead-d925-4058-a390-88426775ebfa\" (UID: \"47be5ead-d925-4058-a390-88426775ebfa\") " Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.099197 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47be5ead-d925-4058-a390-88426775ebfa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "47be5ead-d925-4058-a390-88426775ebfa" (UID: "47be5ead-d925-4058-a390-88426775ebfa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.109104 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47be5ead-d925-4058-a390-88426775ebfa-kube-api-access-lh6vf" (OuterVolumeSpecName: "kube-api-access-lh6vf") pod "47be5ead-d925-4058-a390-88426775ebfa" (UID: "47be5ead-d925-4058-a390-88426775ebfa"). InnerVolumeSpecName "kube-api-access-lh6vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.187308 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"71db2c7f-94ac-455b-9142-c63bb53ed942","Type":"ContainerStarted","Data":"a5eb6f68f861be3ef47148f4e9b5c6f828b9f54ac04547febd2adba19e2c8ea6"} Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.188212 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5tqrl" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.190303 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-c67d-account-create-update-d6zs6" event={"ID":"47be5ead-d925-4058-a390-88426775ebfa","Type":"ContainerDied","Data":"ff4b72097f5fa980a04470e08de4f26a6d0b0ce77a977debbf8fa4729d8b6ef5"} Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.190990 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-c67d-account-create-update-d6zs6" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.191251 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff4b72097f5fa980a04470e08de4f26a6d0b0ce77a977debbf8fa4729d8b6ef5" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.197054 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-edd8-account-create-update-f87t8" event={"ID":"b19ce251-54b7-40f3-bbe0-6afb6064b307","Type":"ContainerDied","Data":"cdfddfc4cb8804e6123be5a09adae2421022371cb6d3f95765e08c364825f808"} Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.197085 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cdfddfc4cb8804e6123be5a09adae2421022371cb6d3f95765e08c364825f808" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.198611 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47be5ead-d925-4058-a390-88426775ebfa-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.198625 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh6vf\" (UniqueName: \"kubernetes.io/projected/47be5ead-d925-4058-a390-88426775ebfa-kube-api-access-lh6vf\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.200659 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-edd8-account-create-update-f87t8" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.201208 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2fffc8e4-c48c-4f89-bc5d-143c3d96a557","Type":"ContainerStarted","Data":"f838346034485d97935465544d2d02e919a1acb676bd871c60ff12be92ffd150"} Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.214948 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-28cs6" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.219641 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3c08f474-3395-463f-ac2f-559aa97bb021","Type":"ContainerStarted","Data":"cee0b642143481c27ca2739ec7177c802a23a2eb908f54b09109aa2ff1c06404"} Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.237789 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-28cs6" event={"ID":"b544108c-9253-40f1-a5ed-8456173a3e68","Type":"ContainerDied","Data":"6c8ce06fc3243cd513cd787f8c8c245105f1ff74204738c70ce23c131ae91698"} Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.237863 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c8ce06fc3243cd513cd787f8c8c245105f1ff74204738c70ce23c131ae91698" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.237950 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-28cs6" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.241087 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.241070733 podStartE2EDuration="6.241070733s" podCreationTimestamp="2025-12-10 19:21:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:21:19.238726141 +0000 UTC m=+1559.749337146" watchObservedRunningTime="2025-12-10 19:21:19.241070733 +0000 UTC m=+1559.751681738" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.265350 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5tqrl" event={"ID":"d24aa875-f0bd-4e7d-a985-c633a246fb30","Type":"ContainerDied","Data":"f8e3cba33baef3ea5e00d70fa7e64264e5324cd3f5d1e1da9a595c0cbfd31f12"} Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.265390 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8e3cba33baef3ea5e00d70fa7e64264e5324cd3f5d1e1da9a595c0cbfd31f12" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.265443 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5tqrl" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.272015 4828 generic.go:334] "Generic (PLEG): container finished" podID="63fd1651-0e27-40c0-b5a4-38ab90a957ca" containerID="77393e75c6bbabeacbd9eb70da2c06c290ab9c019eb54d67bdd6ac645d94ee31" exitCode=1 Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.272800 4828 scope.go:117] "RemoveContainer" containerID="7761a7fab7f1757991c79aa7082709c281bce53ed2c6d4d94fb4ef159ff329b9" Dec 10 19:21:19 crc kubenswrapper[4828]: E1210 19:21:19.273072 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-6dfc7956d8-nq8qr_openstack(dbfc3632-aaef-44db-a52a-7403cae01ad4)\"" pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" podUID="dbfc3632-aaef-44db-a52a-7403cae01ad4" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.273534 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-566d76fd44-v6jl7" event={"ID":"63fd1651-0e27-40c0-b5a4-38ab90a957ca","Type":"ContainerDied","Data":"77393e75c6bbabeacbd9eb70da2c06c290ab9c019eb54d67bdd6ac645d94ee31"} Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.273700 4828 scope.go:117] "RemoveContainer" containerID="e36067d92c45d3ac2f0eaae0855f770be2bbe334104e3cbe368c63096553a519" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.273811 4828 scope.go:117] "RemoveContainer" containerID="77393e75c6bbabeacbd9eb70da2c06c290ab9c019eb54d67bdd6ac645d94ee31" Dec 10 19:21:19 crc kubenswrapper[4828]: E1210 19:21:19.274085 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-566d76fd44-v6jl7_openstack(63fd1651-0e27-40c0-b5a4-38ab90a957ca)\"" pod="openstack/heat-api-566d76fd44-v6jl7" podUID="63fd1651-0e27-40c0-b5a4-38ab90a957ca" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.303084 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d24aa875-f0bd-4e7d-a985-c633a246fb30-operator-scripts\") pod \"d24aa875-f0bd-4e7d-a985-c633a246fb30\" (UID: \"d24aa875-f0bd-4e7d-a985-c633a246fb30\") " Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.303227 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnb2k\" (UniqueName: \"kubernetes.io/projected/b19ce251-54b7-40f3-bbe0-6afb6064b307-kube-api-access-qnb2k\") pod \"b19ce251-54b7-40f3-bbe0-6afb6064b307\" (UID: \"b19ce251-54b7-40f3-bbe0-6afb6064b307\") " Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.303287 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqzvj\" (UniqueName: \"kubernetes.io/projected/b544108c-9253-40f1-a5ed-8456173a3e68-kube-api-access-bqzvj\") pod \"b544108c-9253-40f1-a5ed-8456173a3e68\" (UID: \"b544108c-9253-40f1-a5ed-8456173a3e68\") " Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.303318 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ft78c\" (UniqueName: \"kubernetes.io/projected/d24aa875-f0bd-4e7d-a985-c633a246fb30-kube-api-access-ft78c\") pod \"d24aa875-f0bd-4e7d-a985-c633a246fb30\" (UID: \"d24aa875-f0bd-4e7d-a985-c633a246fb30\") " Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.303490 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b19ce251-54b7-40f3-bbe0-6afb6064b307-operator-scripts\") pod \"b19ce251-54b7-40f3-bbe0-6afb6064b307\" (UID: \"b19ce251-54b7-40f3-bbe0-6afb6064b307\") " Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.303629 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b544108c-9253-40f1-a5ed-8456173a3e68-operator-scripts\") pod \"b544108c-9253-40f1-a5ed-8456173a3e68\" (UID: \"b544108c-9253-40f1-a5ed-8456173a3e68\") " Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.304238 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b19ce251-54b7-40f3-bbe0-6afb6064b307-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b19ce251-54b7-40f3-bbe0-6afb6064b307" (UID: "b19ce251-54b7-40f3-bbe0-6afb6064b307"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.304322 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d24aa875-f0bd-4e7d-a985-c633a246fb30-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d24aa875-f0bd-4e7d-a985-c633a246fb30" (UID: "d24aa875-f0bd-4e7d-a985-c633a246fb30"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.304672 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b544108c-9253-40f1-a5ed-8456173a3e68-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b544108c-9253-40f1-a5ed-8456173a3e68" (UID: "b544108c-9253-40f1-a5ed-8456173a3e68"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.306952 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.306923227 podStartE2EDuration="6.306923227s" podCreationTimestamp="2025-12-10 19:21:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:21:19.288649233 +0000 UTC m=+1559.799260238" watchObservedRunningTime="2025-12-10 19:21:19.306923227 +0000 UTC m=+1559.817534232" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.308913 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d24aa875-f0bd-4e7d-a985-c633a246fb30-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.308935 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b19ce251-54b7-40f3-bbe0-6afb6064b307-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.308944 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b544108c-9253-40f1-a5ed-8456173a3e68-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.312583 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b19ce251-54b7-40f3-bbe0-6afb6064b307-kube-api-access-qnb2k" (OuterVolumeSpecName: "kube-api-access-qnb2k") pod "b19ce251-54b7-40f3-bbe0-6afb6064b307" (UID: "b19ce251-54b7-40f3-bbe0-6afb6064b307"). InnerVolumeSpecName "kube-api-access-qnb2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.312713 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d24aa875-f0bd-4e7d-a985-c633a246fb30-kube-api-access-ft78c" (OuterVolumeSpecName: "kube-api-access-ft78c") pod "d24aa875-f0bd-4e7d-a985-c633a246fb30" (UID: "d24aa875-f0bd-4e7d-a985-c633a246fb30"). InnerVolumeSpecName "kube-api-access-ft78c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.315047 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b544108c-9253-40f1-a5ed-8456173a3e68-kube-api-access-bqzvj" (OuterVolumeSpecName: "kube-api-access-bqzvj") pod "b544108c-9253-40f1-a5ed-8456173a3e68" (UID: "b544108c-9253-40f1-a5ed-8456173a3e68"). InnerVolumeSpecName "kube-api-access-bqzvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.413468 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnb2k\" (UniqueName: \"kubernetes.io/projected/b19ce251-54b7-40f3-bbe0-6afb6064b307-kube-api-access-qnb2k\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.413499 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqzvj\" (UniqueName: \"kubernetes.io/projected/b544108c-9253-40f1-a5ed-8456173a3e68-kube-api-access-bqzvj\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.413509 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ft78c\" (UniqueName: \"kubernetes.io/projected/d24aa875-f0bd-4e7d-a985-c633a246fb30-kube-api-access-ft78c\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.710749 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="8cb3c5d9-61fb-415a-a644-75803d3f7c8f" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.201:8776/healthcheck\": dial tcp 10.217.0.201:8776: i/o timeout (Client.Timeout exceeded while awaiting headers)" Dec 10 19:21:19 crc kubenswrapper[4828]: I1210 19:21:19.800726 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:21:19 crc kubenswrapper[4828]: E1210 19:21:19.803787 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:21:20 crc kubenswrapper[4828]: I1210 19:21:20.305276 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"71db2c7f-94ac-455b-9142-c63bb53ed942","Type":"ContainerStarted","Data":"73f3c5ac90816b731b3a078e06998705399010d41bc70705690d22d89c169aa7"} Dec 10 19:21:20 crc kubenswrapper[4828]: I1210 19:21:20.313471 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37649bc1-48ed-4cc3-9b7b-788469c5a989","Type":"ContainerStarted","Data":"b3e3b5b09e81ecd76c11a901248d277635449814b883a34bfdde87a9247480c9"} Dec 10 19:21:20 crc kubenswrapper[4828]: I1210 19:21:20.313599 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerName="ceilometer-central-agent" containerID="cri-o://1dfa7423ff1d5a745b35b1ffa2137f0fe17c49f0bfaec5617f8817a60827ac18" gracePeriod=30 Dec 10 19:21:20 crc kubenswrapper[4828]: I1210 19:21:20.313773 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerName="proxy-httpd" containerID="cri-o://b3e3b5b09e81ecd76c11a901248d277635449814b883a34bfdde87a9247480c9" gracePeriod=30 Dec 10 19:21:20 crc kubenswrapper[4828]: I1210 19:21:20.313856 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerName="sg-core" containerID="cri-o://90ce2e9ec5dd43642c4b2efb3407b0bda330694048ed09ab454805247c0c9905" gracePeriod=30 Dec 10 19:21:20 crc kubenswrapper[4828]: I1210 19:21:20.313619 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 19:21:20 crc kubenswrapper[4828]: I1210 19:21:20.313919 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerName="ceilometer-notification-agent" containerID="cri-o://b78a3e2e45e64cfdc98c611229d668709a2ba675192f598624e1f93a72b18fb9" gracePeriod=30 Dec 10 19:21:20 crc kubenswrapper[4828]: I1210 19:21:20.324127 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-edd8-account-create-update-f87t8" Dec 10 19:21:20 crc kubenswrapper[4828]: I1210 19:21:20.325459 4828 scope.go:117] "RemoveContainer" containerID="77393e75c6bbabeacbd9eb70da2c06c290ab9c019eb54d67bdd6ac645d94ee31" Dec 10 19:21:20 crc kubenswrapper[4828]: E1210 19:21:20.325727 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-566d76fd44-v6jl7_openstack(63fd1651-0e27-40c0-b5a4-38ab90a957ca)\"" pod="openstack/heat-api-566d76fd44-v6jl7" podUID="63fd1651-0e27-40c0-b5a4-38ab90a957ca" Dec 10 19:21:20 crc kubenswrapper[4828]: I1210 19:21:20.337703 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=14.026358076 podStartE2EDuration="20.337686306s" podCreationTimestamp="2025-12-10 19:21:00 +0000 UTC" firstStartedPulling="2025-12-10 19:21:12.670645281 +0000 UTC m=+1553.181256286" lastFinishedPulling="2025-12-10 19:21:18.981973511 +0000 UTC m=+1559.492584516" observedRunningTime="2025-12-10 19:21:20.335208851 +0000 UTC m=+1560.845819856" watchObservedRunningTime="2025-12-10 19:21:20.337686306 +0000 UTC m=+1560.848297311" Dec 10 19:21:21 crc kubenswrapper[4828]: I1210 19:21:21.338966 4828 generic.go:334] "Generic (PLEG): container finished" podID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerID="b3e3b5b09e81ecd76c11a901248d277635449814b883a34bfdde87a9247480c9" exitCode=0 Dec 10 19:21:21 crc kubenswrapper[4828]: I1210 19:21:21.339295 4828 generic.go:334] "Generic (PLEG): container finished" podID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerID="90ce2e9ec5dd43642c4b2efb3407b0bda330694048ed09ab454805247c0c9905" exitCode=2 Dec 10 19:21:21 crc kubenswrapper[4828]: I1210 19:21:21.339305 4828 generic.go:334] "Generic (PLEG): container finished" podID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerID="b78a3e2e45e64cfdc98c611229d668709a2ba675192f598624e1f93a72b18fb9" exitCode=0 Dec 10 19:21:21 crc kubenswrapper[4828]: I1210 19:21:21.339064 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37649bc1-48ed-4cc3-9b7b-788469c5a989","Type":"ContainerDied","Data":"b3e3b5b09e81ecd76c11a901248d277635449814b883a34bfdde87a9247480c9"} Dec 10 19:21:21 crc kubenswrapper[4828]: I1210 19:21:21.339371 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37649bc1-48ed-4cc3-9b7b-788469c5a989","Type":"ContainerDied","Data":"90ce2e9ec5dd43642c4b2efb3407b0bda330694048ed09ab454805247c0c9905"} Dec 10 19:21:21 crc kubenswrapper[4828]: I1210 19:21:21.339387 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37649bc1-48ed-4cc3-9b7b-788469c5a989","Type":"ContainerDied","Data":"b78a3e2e45e64cfdc98c611229d668709a2ba675192f598624e1f93a72b18fb9"} Dec 10 19:21:21 crc kubenswrapper[4828]: I1210 19:21:21.341553 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"71db2c7f-94ac-455b-9142-c63bb53ed942","Type":"ContainerStarted","Data":"a65b8fbe8e231bdc5476bb089c6b0ab7f6c314a97a1bcf71907dfbc9fd727d64"} Dec 10 19:21:21 crc kubenswrapper[4828]: I1210 19:21:21.341696 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 10 19:21:21 crc kubenswrapper[4828]: I1210 19:21:21.367186 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.36716366 podStartE2EDuration="4.36716366s" podCreationTimestamp="2025-12-10 19:21:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:21:21.359468377 +0000 UTC m=+1561.870079402" watchObservedRunningTime="2025-12-10 19:21:21.36716366 +0000 UTC m=+1561.877774675" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.029549 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.221262 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37649bc1-48ed-4cc3-9b7b-788469c5a989-log-httpd\") pod \"37649bc1-48ed-4cc3-9b7b-788469c5a989\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.221389 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-combined-ca-bundle\") pod \"37649bc1-48ed-4cc3-9b7b-788469c5a989\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.221425 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-config-data\") pod \"37649bc1-48ed-4cc3-9b7b-788469c5a989\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.221517 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37649bc1-48ed-4cc3-9b7b-788469c5a989-run-httpd\") pod \"37649bc1-48ed-4cc3-9b7b-788469c5a989\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.221607 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-scripts\") pod \"37649bc1-48ed-4cc3-9b7b-788469c5a989\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.221632 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrv4t\" (UniqueName: \"kubernetes.io/projected/37649bc1-48ed-4cc3-9b7b-788469c5a989-kube-api-access-vrv4t\") pod \"37649bc1-48ed-4cc3-9b7b-788469c5a989\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.221680 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-sg-core-conf-yaml\") pod \"37649bc1-48ed-4cc3-9b7b-788469c5a989\" (UID: \"37649bc1-48ed-4cc3-9b7b-788469c5a989\") " Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.222085 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37649bc1-48ed-4cc3-9b7b-788469c5a989-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "37649bc1-48ed-4cc3-9b7b-788469c5a989" (UID: "37649bc1-48ed-4cc3-9b7b-788469c5a989"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.222273 4828 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37649bc1-48ed-4cc3-9b7b-788469c5a989-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.222291 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37649bc1-48ed-4cc3-9b7b-788469c5a989-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "37649bc1-48ed-4cc3-9b7b-788469c5a989" (UID: "37649bc1-48ed-4cc3-9b7b-788469c5a989"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.228916 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37649bc1-48ed-4cc3-9b7b-788469c5a989-kube-api-access-vrv4t" (OuterVolumeSpecName: "kube-api-access-vrv4t") pod "37649bc1-48ed-4cc3-9b7b-788469c5a989" (UID: "37649bc1-48ed-4cc3-9b7b-788469c5a989"). InnerVolumeSpecName "kube-api-access-vrv4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.240045 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-scripts" (OuterVolumeSpecName: "scripts") pod "37649bc1-48ed-4cc3-9b7b-788469c5a989" (UID: "37649bc1-48ed-4cc3-9b7b-788469c5a989"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.308967 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "37649bc1-48ed-4cc3-9b7b-788469c5a989" (UID: "37649bc1-48ed-4cc3-9b7b-788469c5a989"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.325075 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.325108 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrv4t\" (UniqueName: \"kubernetes.io/projected/37649bc1-48ed-4cc3-9b7b-788469c5a989-kube-api-access-vrv4t\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.325119 4828 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.325127 4828 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37649bc1-48ed-4cc3-9b7b-788469c5a989-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.349011 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37649bc1-48ed-4cc3-9b7b-788469c5a989" (UID: "37649bc1-48ed-4cc3-9b7b-788469c5a989"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.357858 4828 generic.go:334] "Generic (PLEG): container finished" podID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerID="1dfa7423ff1d5a745b35b1ffa2137f0fe17c49f0bfaec5617f8817a60827ac18" exitCode=0 Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.357943 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.357952 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37649bc1-48ed-4cc3-9b7b-788469c5a989","Type":"ContainerDied","Data":"1dfa7423ff1d5a745b35b1ffa2137f0fe17c49f0bfaec5617f8817a60827ac18"} Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.358012 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37649bc1-48ed-4cc3-9b7b-788469c5a989","Type":"ContainerDied","Data":"b3d3644115949d315422eb108ab764c35379cde648deadfb6e78b9f250f94a14"} Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.358034 4828 scope.go:117] "RemoveContainer" containerID="b3e3b5b09e81ecd76c11a901248d277635449814b883a34bfdde87a9247480c9" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.386861 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-config-data" (OuterVolumeSpecName: "config-data") pod "37649bc1-48ed-4cc3-9b7b-788469c5a989" (UID: "37649bc1-48ed-4cc3-9b7b-788469c5a989"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.398327 4828 scope.go:117] "RemoveContainer" containerID="90ce2e9ec5dd43642c4b2efb3407b0bda330694048ed09ab454805247c0c9905" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.421509 4828 scope.go:117] "RemoveContainer" containerID="b78a3e2e45e64cfdc98c611229d668709a2ba675192f598624e1f93a72b18fb9" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.426616 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.426655 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37649bc1-48ed-4cc3-9b7b-788469c5a989-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.444900 4828 scope.go:117] "RemoveContainer" containerID="1dfa7423ff1d5a745b35b1ffa2137f0fe17c49f0bfaec5617f8817a60827ac18" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.468705 4828 scope.go:117] "RemoveContainer" containerID="b3e3b5b09e81ecd76c11a901248d277635449814b883a34bfdde87a9247480c9" Dec 10 19:21:22 crc kubenswrapper[4828]: E1210 19:21:22.469200 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3e3b5b09e81ecd76c11a901248d277635449814b883a34bfdde87a9247480c9\": container with ID starting with b3e3b5b09e81ecd76c11a901248d277635449814b883a34bfdde87a9247480c9 not found: ID does not exist" containerID="b3e3b5b09e81ecd76c11a901248d277635449814b883a34bfdde87a9247480c9" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.469240 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3e3b5b09e81ecd76c11a901248d277635449814b883a34bfdde87a9247480c9"} err="failed to get container status \"b3e3b5b09e81ecd76c11a901248d277635449814b883a34bfdde87a9247480c9\": rpc error: code = NotFound desc = could not find container \"b3e3b5b09e81ecd76c11a901248d277635449814b883a34bfdde87a9247480c9\": container with ID starting with b3e3b5b09e81ecd76c11a901248d277635449814b883a34bfdde87a9247480c9 not found: ID does not exist" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.469264 4828 scope.go:117] "RemoveContainer" containerID="90ce2e9ec5dd43642c4b2efb3407b0bda330694048ed09ab454805247c0c9905" Dec 10 19:21:22 crc kubenswrapper[4828]: E1210 19:21:22.469657 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90ce2e9ec5dd43642c4b2efb3407b0bda330694048ed09ab454805247c0c9905\": container with ID starting with 90ce2e9ec5dd43642c4b2efb3407b0bda330694048ed09ab454805247c0c9905 not found: ID does not exist" containerID="90ce2e9ec5dd43642c4b2efb3407b0bda330694048ed09ab454805247c0c9905" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.469679 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90ce2e9ec5dd43642c4b2efb3407b0bda330694048ed09ab454805247c0c9905"} err="failed to get container status \"90ce2e9ec5dd43642c4b2efb3407b0bda330694048ed09ab454805247c0c9905\": rpc error: code = NotFound desc = could not find container \"90ce2e9ec5dd43642c4b2efb3407b0bda330694048ed09ab454805247c0c9905\": container with ID starting with 90ce2e9ec5dd43642c4b2efb3407b0bda330694048ed09ab454805247c0c9905 not found: ID does not exist" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.469693 4828 scope.go:117] "RemoveContainer" containerID="b78a3e2e45e64cfdc98c611229d668709a2ba675192f598624e1f93a72b18fb9" Dec 10 19:21:22 crc kubenswrapper[4828]: E1210 19:21:22.470057 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b78a3e2e45e64cfdc98c611229d668709a2ba675192f598624e1f93a72b18fb9\": container with ID starting with b78a3e2e45e64cfdc98c611229d668709a2ba675192f598624e1f93a72b18fb9 not found: ID does not exist" containerID="b78a3e2e45e64cfdc98c611229d668709a2ba675192f598624e1f93a72b18fb9" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.470082 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b78a3e2e45e64cfdc98c611229d668709a2ba675192f598624e1f93a72b18fb9"} err="failed to get container status \"b78a3e2e45e64cfdc98c611229d668709a2ba675192f598624e1f93a72b18fb9\": rpc error: code = NotFound desc = could not find container \"b78a3e2e45e64cfdc98c611229d668709a2ba675192f598624e1f93a72b18fb9\": container with ID starting with b78a3e2e45e64cfdc98c611229d668709a2ba675192f598624e1f93a72b18fb9 not found: ID does not exist" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.470097 4828 scope.go:117] "RemoveContainer" containerID="1dfa7423ff1d5a745b35b1ffa2137f0fe17c49f0bfaec5617f8817a60827ac18" Dec 10 19:21:22 crc kubenswrapper[4828]: E1210 19:21:22.470387 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dfa7423ff1d5a745b35b1ffa2137f0fe17c49f0bfaec5617f8817a60827ac18\": container with ID starting with 1dfa7423ff1d5a745b35b1ffa2137f0fe17c49f0bfaec5617f8817a60827ac18 not found: ID does not exist" containerID="1dfa7423ff1d5a745b35b1ffa2137f0fe17c49f0bfaec5617f8817a60827ac18" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.470431 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dfa7423ff1d5a745b35b1ffa2137f0fe17c49f0bfaec5617f8817a60827ac18"} err="failed to get container status \"1dfa7423ff1d5a745b35b1ffa2137f0fe17c49f0bfaec5617f8817a60827ac18\": rpc error: code = NotFound desc = could not find container \"1dfa7423ff1d5a745b35b1ffa2137f0fe17c49f0bfaec5617f8817a60827ac18\": container with ID starting with 1dfa7423ff1d5a745b35b1ffa2137f0fe17c49f0bfaec5617f8817a60827ac18 not found: ID does not exist" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.720154 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.731664 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.745401 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:21:22 crc kubenswrapper[4828]: E1210 19:21:22.745938 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b19ce251-54b7-40f3-bbe0-6afb6064b307" containerName="mariadb-account-create-update" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.745955 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="b19ce251-54b7-40f3-bbe0-6afb6064b307" containerName="mariadb-account-create-update" Dec 10 19:21:22 crc kubenswrapper[4828]: E1210 19:21:22.745970 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerName="proxy-httpd" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.745977 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerName="proxy-httpd" Dec 10 19:21:22 crc kubenswrapper[4828]: E1210 19:21:22.746006 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b544108c-9253-40f1-a5ed-8456173a3e68" containerName="mariadb-database-create" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.746013 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="b544108c-9253-40f1-a5ed-8456173a3e68" containerName="mariadb-database-create" Dec 10 19:21:22 crc kubenswrapper[4828]: E1210 19:21:22.746030 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d24aa875-f0bd-4e7d-a985-c633a246fb30" containerName="mariadb-database-create" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.746039 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="d24aa875-f0bd-4e7d-a985-c633a246fb30" containerName="mariadb-database-create" Dec 10 19:21:22 crc kubenswrapper[4828]: E1210 19:21:22.746056 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerName="ceilometer-notification-agent" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.746062 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerName="ceilometer-notification-agent" Dec 10 19:21:22 crc kubenswrapper[4828]: E1210 19:21:22.746076 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f5c519b-682a-430f-b1ed-ee93f6b7c9f2" containerName="mariadb-account-create-update" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.746082 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f5c519b-682a-430f-b1ed-ee93f6b7c9f2" containerName="mariadb-account-create-update" Dec 10 19:21:22 crc kubenswrapper[4828]: E1210 19:21:22.746094 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47be5ead-d925-4058-a390-88426775ebfa" containerName="mariadb-account-create-update" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.746099 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="47be5ead-d925-4058-a390-88426775ebfa" containerName="mariadb-account-create-update" Dec 10 19:21:22 crc kubenswrapper[4828]: E1210 19:21:22.746111 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerName="sg-core" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.746116 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerName="sg-core" Dec 10 19:21:22 crc kubenswrapper[4828]: E1210 19:21:22.746128 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerName="ceilometer-central-agent" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.746134 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerName="ceilometer-central-agent" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.746322 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="d24aa875-f0bd-4e7d-a985-c633a246fb30" containerName="mariadb-database-create" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.746343 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerName="ceilometer-central-agent" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.746355 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f5c519b-682a-430f-b1ed-ee93f6b7c9f2" containerName="mariadb-account-create-update" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.746366 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerName="proxy-httpd" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.746380 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="b544108c-9253-40f1-a5ed-8456173a3e68" containerName="mariadb-database-create" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.746389 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="47be5ead-d925-4058-a390-88426775ebfa" containerName="mariadb-account-create-update" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.746398 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerName="sg-core" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.746411 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="37649bc1-48ed-4cc3-9b7b-788469c5a989" containerName="ceilometer-notification-agent" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.746419 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="b19ce251-54b7-40f3-bbe0-6afb6064b307" containerName="mariadb-account-create-update" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.748344 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.750773 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.756566 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.761878 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.943477 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.943536 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-config-data\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.943720 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-run-httpd\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.943835 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.943855 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-scripts\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.944061 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gjcx\" (UniqueName: \"kubernetes.io/projected/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-kube-api-access-4gjcx\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:22 crc kubenswrapper[4828]: I1210 19:21:22.944147 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-log-httpd\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.047126 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.047170 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-config-data\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.047278 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-run-httpd\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.047307 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.047323 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-scripts\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.047395 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gjcx\" (UniqueName: \"kubernetes.io/projected/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-kube-api-access-4gjcx\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.047428 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-log-httpd\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.047888 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-log-httpd\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.047896 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-run-httpd\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.052030 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-scripts\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.052145 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.052252 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.062821 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-config-data\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.072257 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gjcx\" (UniqueName: \"kubernetes.io/projected/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-kube-api-access-4gjcx\") pod \"ceilometer-0\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " pod="openstack/ceilometer-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.364679 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.451566 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.451616 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.505407 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.513769 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.666084 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.710954 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.711717 4828 scope.go:117] "RemoveContainer" containerID="77393e75c6bbabeacbd9eb70da2c06c290ab9c019eb54d67bdd6ac645d94ee31" Dec 10 19:21:23 crc kubenswrapper[4828]: E1210 19:21:23.711997 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-566d76fd44-v6jl7_openstack(63fd1651-0e27-40c0-b5a4-38ab90a957ca)\"" pod="openstack/heat-api-566d76fd44-v6jl7" podUID="63fd1651-0e27-40c0-b5a4-38ab90a957ca" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.712044 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.724827 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-566d76fd44-v6jl7"] Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.802491 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37649bc1-48ed-4cc3-9b7b-788469c5a989" path="/var/lib/kubelet/pods/37649bc1-48ed-4cc3-9b7b-788469c5a989/volumes" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.853602 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.885952 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:21:23 crc kubenswrapper[4828]: I1210 19:21:23.943287 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6dfc7956d8-nq8qr"] Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.387902 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" event={"ID":"dbfc3632-aaef-44db-a52a-7403cae01ad4","Type":"ContainerDied","Data":"88ccd8b0d00966fb6e4dadab1cc9c578a434c80e53a522a22b1f66ab116dfbe2"} Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.388980 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88ccd8b0d00966fb6e4dadab1cc9c578a434c80e53a522a22b1f66ab116dfbe2" Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.398584 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f8ce74b-4f95-4e0b-b073-15450cbb4f90","Type":"ContainerStarted","Data":"e105479d91abf117658c10272ed832817771ef1b1953fea583869c5e26dd6754"} Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.398878 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.399366 4828 scope.go:117] "RemoveContainer" containerID="77393e75c6bbabeacbd9eb70da2c06c290ab9c019eb54d67bdd6ac645d94ee31" Dec 10 19:21:24 crc kubenswrapper[4828]: E1210 19:21:24.399721 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-566d76fd44-v6jl7_openstack(63fd1651-0e27-40c0-b5a4-38ab90a957ca)\"" pod="openstack/heat-api-566d76fd44-v6jl7" podUID="63fd1651-0e27-40c0-b5a4-38ab90a957ca" Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.400142 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.504826 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.585050 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85rsj\" (UniqueName: \"kubernetes.io/projected/dbfc3632-aaef-44db-a52a-7403cae01ad4-kube-api-access-85rsj\") pod \"dbfc3632-aaef-44db-a52a-7403cae01ad4\" (UID: \"dbfc3632-aaef-44db-a52a-7403cae01ad4\") " Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.585139 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-combined-ca-bundle\") pod \"dbfc3632-aaef-44db-a52a-7403cae01ad4\" (UID: \"dbfc3632-aaef-44db-a52a-7403cae01ad4\") " Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.585162 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-config-data\") pod \"dbfc3632-aaef-44db-a52a-7403cae01ad4\" (UID: \"dbfc3632-aaef-44db-a52a-7403cae01ad4\") " Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.585320 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-config-data-custom\") pod \"dbfc3632-aaef-44db-a52a-7403cae01ad4\" (UID: \"dbfc3632-aaef-44db-a52a-7403cae01ad4\") " Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.595340 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbfc3632-aaef-44db-a52a-7403cae01ad4-kube-api-access-85rsj" (OuterVolumeSpecName: "kube-api-access-85rsj") pod "dbfc3632-aaef-44db-a52a-7403cae01ad4" (UID: "dbfc3632-aaef-44db-a52a-7403cae01ad4"). InnerVolumeSpecName "kube-api-access-85rsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.599175 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "dbfc3632-aaef-44db-a52a-7403cae01ad4" (UID: "dbfc3632-aaef-44db-a52a-7403cae01ad4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.623020 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dbfc3632-aaef-44db-a52a-7403cae01ad4" (UID: "dbfc3632-aaef-44db-a52a-7403cae01ad4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.656936 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-config-data" (OuterVolumeSpecName: "config-data") pod "dbfc3632-aaef-44db-a52a-7403cae01ad4" (UID: "dbfc3632-aaef-44db-a52a-7403cae01ad4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.689037 4828 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.689070 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85rsj\" (UniqueName: \"kubernetes.io/projected/dbfc3632-aaef-44db-a52a-7403cae01ad4-kube-api-access-85rsj\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.689105 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:24 crc kubenswrapper[4828]: I1210 19:21:24.689115 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbfc3632-aaef-44db-a52a-7403cae01ad4-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.224832 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.225377 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.267613 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.302279 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.442599 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6dfc7956d8-nq8qr" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.443071 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f8ce74b-4f95-4e0b-b073-15450cbb4f90","Type":"ContainerStarted","Data":"bf8d18b5b59b5b77dcb8260d924335e0772a03f59825ce0445c896b38f70f955"} Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.443697 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.443720 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.590385 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-6dfc7956d8-nq8qr"] Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.603274 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-6dfc7956d8-nq8qr"] Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.622459 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mzd68"] Dec 10 19:21:25 crc kubenswrapper[4828]: E1210 19:21:25.623095 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbfc3632-aaef-44db-a52a-7403cae01ad4" containerName="heat-cfnapi" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.623115 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbfc3632-aaef-44db-a52a-7403cae01ad4" containerName="heat-cfnapi" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.623336 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbfc3632-aaef-44db-a52a-7403cae01ad4" containerName="heat-cfnapi" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.624192 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mzd68" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.628247 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mzd68"] Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.631291 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-j4mkf" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.631405 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.631597 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.729215 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9d95\" (UniqueName: \"kubernetes.io/projected/ff038e8c-07a0-41ca-9719-3b8e326721a0-kube-api-access-j9d95\") pod \"nova-cell0-conductor-db-sync-mzd68\" (UID: \"ff038e8c-07a0-41ca-9719-3b8e326721a0\") " pod="openstack/nova-cell0-conductor-db-sync-mzd68" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.729298 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-scripts\") pod \"nova-cell0-conductor-db-sync-mzd68\" (UID: \"ff038e8c-07a0-41ca-9719-3b8e326721a0\") " pod="openstack/nova-cell0-conductor-db-sync-mzd68" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.729439 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mzd68\" (UID: \"ff038e8c-07a0-41ca-9719-3b8e326721a0\") " pod="openstack/nova-cell0-conductor-db-sync-mzd68" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.729477 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-config-data\") pod \"nova-cell0-conductor-db-sync-mzd68\" (UID: \"ff038e8c-07a0-41ca-9719-3b8e326721a0\") " pod="openstack/nova-cell0-conductor-db-sync-mzd68" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.803664 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbfc3632-aaef-44db-a52a-7403cae01ad4" path="/var/lib/kubelet/pods/dbfc3632-aaef-44db-a52a-7403cae01ad4/volumes" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.831782 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9d95\" (UniqueName: \"kubernetes.io/projected/ff038e8c-07a0-41ca-9719-3b8e326721a0-kube-api-access-j9d95\") pod \"nova-cell0-conductor-db-sync-mzd68\" (UID: \"ff038e8c-07a0-41ca-9719-3b8e326721a0\") " pod="openstack/nova-cell0-conductor-db-sync-mzd68" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.831849 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-scripts\") pod \"nova-cell0-conductor-db-sync-mzd68\" (UID: \"ff038e8c-07a0-41ca-9719-3b8e326721a0\") " pod="openstack/nova-cell0-conductor-db-sync-mzd68" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.831931 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mzd68\" (UID: \"ff038e8c-07a0-41ca-9719-3b8e326721a0\") " pod="openstack/nova-cell0-conductor-db-sync-mzd68" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.831961 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-config-data\") pod \"nova-cell0-conductor-db-sync-mzd68\" (UID: \"ff038e8c-07a0-41ca-9719-3b8e326721a0\") " pod="openstack/nova-cell0-conductor-db-sync-mzd68" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.839210 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-scripts\") pod \"nova-cell0-conductor-db-sync-mzd68\" (UID: \"ff038e8c-07a0-41ca-9719-3b8e326721a0\") " pod="openstack/nova-cell0-conductor-db-sync-mzd68" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.839348 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-config-data\") pod \"nova-cell0-conductor-db-sync-mzd68\" (UID: \"ff038e8c-07a0-41ca-9719-3b8e326721a0\") " pod="openstack/nova-cell0-conductor-db-sync-mzd68" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.850556 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-mzd68\" (UID: \"ff038e8c-07a0-41ca-9719-3b8e326721a0\") " pod="openstack/nova-cell0-conductor-db-sync-mzd68" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.852124 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9d95\" (UniqueName: \"kubernetes.io/projected/ff038e8c-07a0-41ca-9719-3b8e326721a0-kube-api-access-j9d95\") pod \"nova-cell0-conductor-db-sync-mzd68\" (UID: \"ff038e8c-07a0-41ca-9719-3b8e326721a0\") " pod="openstack/nova-cell0-conductor-db-sync-mzd68" Dec 10 19:21:25 crc kubenswrapper[4828]: I1210 19:21:25.950276 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mzd68" Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.138832 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.247580 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-config-data\") pod \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\" (UID: \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\") " Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.247769 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-config-data-custom\") pod \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\" (UID: \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\") " Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.247839 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jl2q\" (UniqueName: \"kubernetes.io/projected/63fd1651-0e27-40c0-b5a4-38ab90a957ca-kube-api-access-2jl2q\") pod \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\" (UID: \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\") " Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.247922 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-combined-ca-bundle\") pod \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\" (UID: \"63fd1651-0e27-40c0-b5a4-38ab90a957ca\") " Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.269096 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63fd1651-0e27-40c0-b5a4-38ab90a957ca-kube-api-access-2jl2q" (OuterVolumeSpecName: "kube-api-access-2jl2q") pod "63fd1651-0e27-40c0-b5a4-38ab90a957ca" (UID: "63fd1651-0e27-40c0-b5a4-38ab90a957ca"). InnerVolumeSpecName "kube-api-access-2jl2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.269234 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "63fd1651-0e27-40c0-b5a4-38ab90a957ca" (UID: "63fd1651-0e27-40c0-b5a4-38ab90a957ca"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.328915 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63fd1651-0e27-40c0-b5a4-38ab90a957ca" (UID: "63fd1651-0e27-40c0-b5a4-38ab90a957ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.350955 4828 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.350994 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jl2q\" (UniqueName: \"kubernetes.io/projected/63fd1651-0e27-40c0-b5a4-38ab90a957ca-kube-api-access-2jl2q\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.351008 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.439926 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-config-data" (OuterVolumeSpecName: "config-data") pod "63fd1651-0e27-40c0-b5a4-38ab90a957ca" (UID: "63fd1651-0e27-40c0-b5a4-38ab90a957ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.454927 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63fd1651-0e27-40c0-b5a4-38ab90a957ca-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.481649 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-566d76fd44-v6jl7" event={"ID":"63fd1651-0e27-40c0-b5a4-38ab90a957ca","Type":"ContainerDied","Data":"816072689e5886b9ace475f1dd2b93dc9cad209a730993c27633571a5d189cc6"} Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.481703 4828 scope.go:117] "RemoveContainer" containerID="77393e75c6bbabeacbd9eb70da2c06c290ab9c019eb54d67bdd6ac645d94ee31" Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.481880 4828 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.481891 4828 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.482598 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-566d76fd44-v6jl7" Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.555296 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-566d76fd44-v6jl7"] Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.566953 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-566d76fd44-v6jl7"] Dec 10 19:21:26 crc kubenswrapper[4828]: I1210 19:21:26.811652 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mzd68"] Dec 10 19:21:26 crc kubenswrapper[4828]: E1210 19:21:26.924493 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice/crio-cdfddfc4cb8804e6123be5a09adae2421022371cb6d3f95765e08c364825f808\": RecentStats: unable to find data in memory cache]" Dec 10 19:21:27 crc kubenswrapper[4828]: I1210 19:21:27.495067 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mzd68" event={"ID":"ff038e8c-07a0-41ca-9719-3b8e326721a0","Type":"ContainerStarted","Data":"90a30594231e75e9ca3bdbd4f0801e306e052122be51be16baa3c1c3f41508da"} Dec 10 19:21:27 crc kubenswrapper[4828]: I1210 19:21:27.498874 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f8ce74b-4f95-4e0b-b073-15450cbb4f90","Type":"ContainerStarted","Data":"ee611da9ecabef08e3128b639dd291809c76e57977a77ac75e4bdacd40a37d99"} Dec 10 19:21:27 crc kubenswrapper[4828]: I1210 19:21:27.814415 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63fd1651-0e27-40c0-b5a4-38ab90a957ca" path="/var/lib/kubelet/pods/63fd1651-0e27-40c0-b5a4-38ab90a957ca/volumes" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.089581 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k6jt6"] Dec 10 19:21:28 crc kubenswrapper[4828]: E1210 19:21:28.090303 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63fd1651-0e27-40c0-b5a4-38ab90a957ca" containerName="heat-api" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.090319 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="63fd1651-0e27-40c0-b5a4-38ab90a957ca" containerName="heat-api" Dec 10 19:21:28 crc kubenswrapper[4828]: E1210 19:21:28.090344 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63fd1651-0e27-40c0-b5a4-38ab90a957ca" containerName="heat-api" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.090351 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="63fd1651-0e27-40c0-b5a4-38ab90a957ca" containerName="heat-api" Dec 10 19:21:28 crc kubenswrapper[4828]: E1210 19:21:28.090382 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbfc3632-aaef-44db-a52a-7403cae01ad4" containerName="heat-cfnapi" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.090388 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbfc3632-aaef-44db-a52a-7403cae01ad4" containerName="heat-cfnapi" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.090582 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbfc3632-aaef-44db-a52a-7403cae01ad4" containerName="heat-cfnapi" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.090611 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="63fd1651-0e27-40c0-b5a4-38ab90a957ca" containerName="heat-api" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.090620 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="63fd1651-0e27-40c0-b5a4-38ab90a957ca" containerName="heat-api" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.094084 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k6jt6" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.162951 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k6jt6"] Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.203757 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8840b2c3-1f04-4623-8108-eb86d16589cc-utilities\") pod \"certified-operators-k6jt6\" (UID: \"8840b2c3-1f04-4623-8108-eb86d16589cc\") " pod="openshift-marketplace/certified-operators-k6jt6" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.203891 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zmsn\" (UniqueName: \"kubernetes.io/projected/8840b2c3-1f04-4623-8108-eb86d16589cc-kube-api-access-8zmsn\") pod \"certified-operators-k6jt6\" (UID: \"8840b2c3-1f04-4623-8108-eb86d16589cc\") " pod="openshift-marketplace/certified-operators-k6jt6" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.204109 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8840b2c3-1f04-4623-8108-eb86d16589cc-catalog-content\") pod \"certified-operators-k6jt6\" (UID: \"8840b2c3-1f04-4623-8108-eb86d16589cc\") " pod="openshift-marketplace/certified-operators-k6jt6" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.306051 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8840b2c3-1f04-4623-8108-eb86d16589cc-catalog-content\") pod \"certified-operators-k6jt6\" (UID: \"8840b2c3-1f04-4623-8108-eb86d16589cc\") " pod="openshift-marketplace/certified-operators-k6jt6" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.306153 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8840b2c3-1f04-4623-8108-eb86d16589cc-utilities\") pod \"certified-operators-k6jt6\" (UID: \"8840b2c3-1f04-4623-8108-eb86d16589cc\") " pod="openshift-marketplace/certified-operators-k6jt6" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.306212 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zmsn\" (UniqueName: \"kubernetes.io/projected/8840b2c3-1f04-4623-8108-eb86d16589cc-kube-api-access-8zmsn\") pod \"certified-operators-k6jt6\" (UID: \"8840b2c3-1f04-4623-8108-eb86d16589cc\") " pod="openshift-marketplace/certified-operators-k6jt6" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.306610 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8840b2c3-1f04-4623-8108-eb86d16589cc-catalog-content\") pod \"certified-operators-k6jt6\" (UID: \"8840b2c3-1f04-4623-8108-eb86d16589cc\") " pod="openshift-marketplace/certified-operators-k6jt6" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.306623 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8840b2c3-1f04-4623-8108-eb86d16589cc-utilities\") pod \"certified-operators-k6jt6\" (UID: \"8840b2c3-1f04-4623-8108-eb86d16589cc\") " pod="openshift-marketplace/certified-operators-k6jt6" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.384474 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zmsn\" (UniqueName: \"kubernetes.io/projected/8840b2c3-1f04-4623-8108-eb86d16589cc-kube-api-access-8zmsn\") pod \"certified-operators-k6jt6\" (UID: \"8840b2c3-1f04-4623-8108-eb86d16589cc\") " pod="openshift-marketplace/certified-operators-k6jt6" Dec 10 19:21:28 crc kubenswrapper[4828]: E1210 19:21:28.405136 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice/crio-cdfddfc4cb8804e6123be5a09adae2421022371cb6d3f95765e08c364825f808\": RecentStats: unable to find data in memory cache]" Dec 10 19:21:28 crc kubenswrapper[4828]: I1210 19:21:28.415334 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k6jt6" Dec 10 19:21:29 crc kubenswrapper[4828]: I1210 19:21:29.575383 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f8ce74b-4f95-4e0b-b073-15450cbb4f90","Type":"ContainerStarted","Data":"8337a232f6d54e5ff0abb55e569be212d7dc79837047c6d1205ea4ee6dbb6686"} Dec 10 19:21:29 crc kubenswrapper[4828]: I1210 19:21:29.626782 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 19:21:29 crc kubenswrapper[4828]: I1210 19:21:29.626906 4828 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 19:21:29 crc kubenswrapper[4828]: I1210 19:21:29.631079 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 10 19:21:29 crc kubenswrapper[4828]: I1210 19:21:29.852839 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k6jt6"] Dec 10 19:21:30 crc kubenswrapper[4828]: I1210 19:21:30.592098 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6jt6" event={"ID":"8840b2c3-1f04-4623-8108-eb86d16589cc","Type":"ContainerStarted","Data":"bd21b7a12881ce8d8c4b33bf4974ded2fb3583fa1ac6fe17e5864f9cd8270974"} Dec 10 19:21:31 crc kubenswrapper[4828]: I1210 19:21:31.599908 4828 generic.go:334] "Generic (PLEG): container finished" podID="8840b2c3-1f04-4623-8108-eb86d16589cc" containerID="7637cb5284a5795603af03d3334931232414b132a7807dc97a95d16ece9e8e8a" exitCode=0 Dec 10 19:21:31 crc kubenswrapper[4828]: I1210 19:21:31.599974 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6jt6" event={"ID":"8840b2c3-1f04-4623-8108-eb86d16589cc","Type":"ContainerDied","Data":"7637cb5284a5795603af03d3334931232414b132a7807dc97a95d16ece9e8e8a"} Dec 10 19:21:31 crc kubenswrapper[4828]: I1210 19:21:31.794262 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:21:31 crc kubenswrapper[4828]: E1210 19:21:31.794514 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:21:31 crc kubenswrapper[4828]: I1210 19:21:31.842156 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="71db2c7f-94ac-455b-9142-c63bb53ed942" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.224:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 19:21:32 crc kubenswrapper[4828]: I1210 19:21:32.066750 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 19:21:32 crc kubenswrapper[4828]: I1210 19:21:32.066903 4828 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 10 19:21:32 crc kubenswrapper[4828]: I1210 19:21:32.068431 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 10 19:21:32 crc kubenswrapper[4828]: I1210 19:21:32.819043 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="71db2c7f-94ac-455b-9142-c63bb53ed942" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.224:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 19:21:34 crc kubenswrapper[4828]: I1210 19:21:34.188473 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:21:34 crc kubenswrapper[4828]: I1210 19:21:34.273033 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-787d69597b-vd29s"] Dec 10 19:21:34 crc kubenswrapper[4828]: I1210 19:21:34.273266 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-787d69597b-vd29s" podUID="abb273e9-4472-4b35-bd92-5c00f7613bd7" containerName="heat-engine" containerID="cri-o://fc35b017bc5e9c9d4621c6f6073a56b6b382e06001a2b8fcd5b8680ffd70a2d5" gracePeriod=60 Dec 10 19:21:34 crc kubenswrapper[4828]: I1210 19:21:34.645296 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f8ce74b-4f95-4e0b-b073-15450cbb4f90","Type":"ContainerStarted","Data":"13b30f8eac0d258f61058ba5030d26c003c6447a783aeaf1a92ad703a07b016e"} Dec 10 19:21:34 crc kubenswrapper[4828]: I1210 19:21:34.645832 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 19:21:34 crc kubenswrapper[4828]: I1210 19:21:34.669684 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=5.61730531 podStartE2EDuration="12.669669126s" podCreationTimestamp="2025-12-10 19:21:22 +0000 UTC" firstStartedPulling="2025-12-10 19:21:23.909755849 +0000 UTC m=+1564.420366854" lastFinishedPulling="2025-12-10 19:21:30.962119665 +0000 UTC m=+1571.472730670" observedRunningTime="2025-12-10 19:21:34.66338238 +0000 UTC m=+1575.173993385" watchObservedRunningTime="2025-12-10 19:21:34.669669126 +0000 UTC m=+1575.180280121" Dec 10 19:21:35 crc kubenswrapper[4828]: E1210 19:21:35.342851 4828 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fc35b017bc5e9c9d4621c6f6073a56b6b382e06001a2b8fcd5b8680ffd70a2d5" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 10 19:21:35 crc kubenswrapper[4828]: E1210 19:21:35.344555 4828 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fc35b017bc5e9c9d4621c6f6073a56b6b382e06001a2b8fcd5b8680ffd70a2d5" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 10 19:21:35 crc kubenswrapper[4828]: E1210 19:21:35.347329 4828 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fc35b017bc5e9c9d4621c6f6073a56b6b382e06001a2b8fcd5b8680ffd70a2d5" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 10 19:21:35 crc kubenswrapper[4828]: E1210 19:21:35.347406 4828 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-787d69597b-vd29s" podUID="abb273e9-4472-4b35-bd92-5c00f7613bd7" containerName="heat-engine" Dec 10 19:21:36 crc kubenswrapper[4828]: I1210 19:21:36.453679 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 10 19:21:38 crc kubenswrapper[4828]: E1210 19:21:38.850872 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice/crio-cdfddfc4cb8804e6123be5a09adae2421022371cb6d3f95765e08c364825f808\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice\": RecentStats: unable to find data in memory cache]" Dec 10 19:21:39 crc kubenswrapper[4828]: I1210 19:21:39.072367 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:21:39 crc kubenswrapper[4828]: I1210 19:21:39.072604 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerName="ceilometer-central-agent" containerID="cri-o://bf8d18b5b59b5b77dcb8260d924335e0772a03f59825ce0445c896b38f70f955" gracePeriod=30 Dec 10 19:21:39 crc kubenswrapper[4828]: I1210 19:21:39.079909 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerName="proxy-httpd" containerID="cri-o://13b30f8eac0d258f61058ba5030d26c003c6447a783aeaf1a92ad703a07b016e" gracePeriod=30 Dec 10 19:21:39 crc kubenswrapper[4828]: I1210 19:21:39.080036 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerName="sg-core" containerID="cri-o://8337a232f6d54e5ff0abb55e569be212d7dc79837047c6d1205ea4ee6dbb6686" gracePeriod=30 Dec 10 19:21:39 crc kubenswrapper[4828]: I1210 19:21:39.080076 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerName="ceilometer-notification-agent" containerID="cri-o://ee611da9ecabef08e3128b639dd291809c76e57977a77ac75e4bdacd40a37d99" gracePeriod=30 Dec 10 19:21:39 crc kubenswrapper[4828]: I1210 19:21:39.732274 4828 generic.go:334] "Generic (PLEG): container finished" podID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerID="13b30f8eac0d258f61058ba5030d26c003c6447a783aeaf1a92ad703a07b016e" exitCode=0 Dec 10 19:21:39 crc kubenswrapper[4828]: I1210 19:21:39.732532 4828 generic.go:334] "Generic (PLEG): container finished" podID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerID="8337a232f6d54e5ff0abb55e569be212d7dc79837047c6d1205ea4ee6dbb6686" exitCode=2 Dec 10 19:21:39 crc kubenswrapper[4828]: I1210 19:21:39.732318 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f8ce74b-4f95-4e0b-b073-15450cbb4f90","Type":"ContainerDied","Data":"13b30f8eac0d258f61058ba5030d26c003c6447a783aeaf1a92ad703a07b016e"} Dec 10 19:21:39 crc kubenswrapper[4828]: I1210 19:21:39.732566 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f8ce74b-4f95-4e0b-b073-15450cbb4f90","Type":"ContainerDied","Data":"8337a232f6d54e5ff0abb55e569be212d7dc79837047c6d1205ea4ee6dbb6686"} Dec 10 19:21:40 crc kubenswrapper[4828]: I1210 19:21:40.746429 4828 generic.go:334] "Generic (PLEG): container finished" podID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerID="ee611da9ecabef08e3128b639dd291809c76e57977a77ac75e4bdacd40a37d99" exitCode=0 Dec 10 19:21:40 crc kubenswrapper[4828]: I1210 19:21:40.746474 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f8ce74b-4f95-4e0b-b073-15450cbb4f90","Type":"ContainerDied","Data":"ee611da9ecabef08e3128b639dd291809c76e57977a77ac75e4bdacd40a37d99"} Dec 10 19:21:41 crc kubenswrapper[4828]: E1210 19:21:41.680266 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice/crio-cdfddfc4cb8804e6123be5a09adae2421022371cb6d3f95765e08c364825f808\": RecentStats: unable to find data in memory cache]" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.017881 4828 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","pod5f9f6cc8-f9aa-478c-9694-bcc05679ba3a"] err="unable to destroy cgroup paths for cgroup [kubepods burstable pod5f9f6cc8-f9aa-478c-9694-bcc05679ba3a] : Timed out while waiting for systemd to remove kubepods-burstable-pod5f9f6cc8_f9aa_478c_9694_bcc05679ba3a.slice" Dec 10 19:21:42 crc kubenswrapper[4828]: E1210 19:21:42.018227 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods burstable pod5f9f6cc8-f9aa-478c-9694-bcc05679ba3a] : unable to destroy cgroup paths for cgroup [kubepods burstable pod5f9f6cc8-f9aa-478c-9694-bcc05679ba3a] : Timed out while waiting for systemd to remove kubepods-burstable-pod5f9f6cc8_f9aa_478c_9694_bcc05679ba3a.slice" pod="openshift-marketplace/redhat-operators-xmnbf" podUID="5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.313614 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.492429 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-config-data\") pod \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.492573 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-scripts\") pod \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.492624 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-sg-core-conf-yaml\") pod \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.492654 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gjcx\" (UniqueName: \"kubernetes.io/projected/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-kube-api-access-4gjcx\") pod \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.492772 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-combined-ca-bundle\") pod \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.492935 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-run-httpd\") pod \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.492964 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-log-httpd\") pod \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\" (UID: \"4f8ce74b-4f95-4e0b-b073-15450cbb4f90\") " Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.493530 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4f8ce74b-4f95-4e0b-b073-15450cbb4f90" (UID: "4f8ce74b-4f95-4e0b-b073-15450cbb4f90"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.493976 4828 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.494058 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4f8ce74b-4f95-4e0b-b073-15450cbb4f90" (UID: "4f8ce74b-4f95-4e0b-b073-15450cbb4f90"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.498740 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-scripts" (OuterVolumeSpecName: "scripts") pod "4f8ce74b-4f95-4e0b-b073-15450cbb4f90" (UID: "4f8ce74b-4f95-4e0b-b073-15450cbb4f90"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.498971 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-kube-api-access-4gjcx" (OuterVolumeSpecName: "kube-api-access-4gjcx") pod "4f8ce74b-4f95-4e0b-b073-15450cbb4f90" (UID: "4f8ce74b-4f95-4e0b-b073-15450cbb4f90"). InnerVolumeSpecName "kube-api-access-4gjcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.526930 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4f8ce74b-4f95-4e0b-b073-15450cbb4f90" (UID: "4f8ce74b-4f95-4e0b-b073-15450cbb4f90"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.596668 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.596706 4828 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.596718 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gjcx\" (UniqueName: \"kubernetes.io/projected/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-kube-api-access-4gjcx\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.596729 4828 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.611289 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f8ce74b-4f95-4e0b-b073-15450cbb4f90" (UID: "4f8ce74b-4f95-4e0b-b073-15450cbb4f90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.672900 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-config-data" (OuterVolumeSpecName: "config-data") pod "4f8ce74b-4f95-4e0b-b073-15450cbb4f90" (UID: "4f8ce74b-4f95-4e0b-b073-15450cbb4f90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.699197 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.699239 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f8ce74b-4f95-4e0b-b073-15450cbb4f90-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.776999 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mzd68" event={"ID":"ff038e8c-07a0-41ca-9719-3b8e326721a0","Type":"ContainerStarted","Data":"c0129558dfdd3e4177e968047ac2960485d6ab7fb4131ec5025c1a85d544ca18"} Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.779659 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6jt6" event={"ID":"8840b2c3-1f04-4623-8108-eb86d16589cc","Type":"ContainerStarted","Data":"0865111c90502cef9b5f495afd3a3834efc58a21e1ff25f1533ad3ebcced5059"} Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.782571 4828 generic.go:334] "Generic (PLEG): container finished" podID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerID="bf8d18b5b59b5b77dcb8260d924335e0772a03f59825ce0445c896b38f70f955" exitCode=0 Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.782605 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f8ce74b-4f95-4e0b-b073-15450cbb4f90","Type":"ContainerDied","Data":"bf8d18b5b59b5b77dcb8260d924335e0772a03f59825ce0445c896b38f70f955"} Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.782640 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f8ce74b-4f95-4e0b-b073-15450cbb4f90","Type":"ContainerDied","Data":"e105479d91abf117658c10272ed832817771ef1b1953fea583869c5e26dd6754"} Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.782654 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xmnbf" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.782675 4828 scope.go:117] "RemoveContainer" containerID="13b30f8eac0d258f61058ba5030d26c003c6447a783aeaf1a92ad703a07b016e" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.782679 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:21:42 crc kubenswrapper[4828]: I1210 19:21:42.812455 4828 scope.go:117] "RemoveContainer" containerID="8337a232f6d54e5ff0abb55e569be212d7dc79837047c6d1205ea4ee6dbb6686" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.027248 4828 scope.go:117] "RemoveContainer" containerID="ee611da9ecabef08e3128b639dd291809c76e57977a77ac75e4bdacd40a37d99" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.036464 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-mzd68" podStartSLOduration=3.095302108 podStartE2EDuration="18.036444952s" podCreationTimestamp="2025-12-10 19:21:25 +0000 UTC" firstStartedPulling="2025-12-10 19:21:26.81115579 +0000 UTC m=+1567.321766785" lastFinishedPulling="2025-12-10 19:21:41.752298624 +0000 UTC m=+1582.262909629" observedRunningTime="2025-12-10 19:21:43.02841883 +0000 UTC m=+1583.539029835" watchObservedRunningTime="2025-12-10 19:21:43.036444952 +0000 UTC m=+1583.547055957" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.055623 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.069831 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.091585 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xmnbf"] Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.105485 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xmnbf"] Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.207936 4828 scope.go:117] "RemoveContainer" containerID="bf8d18b5b59b5b77dcb8260d924335e0772a03f59825ce0445c896b38f70f955" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.232619 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:21:43 crc kubenswrapper[4828]: E1210 19:21:43.233362 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerName="ceilometer-central-agent" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.233386 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerName="ceilometer-central-agent" Dec 10 19:21:43 crc kubenswrapper[4828]: E1210 19:21:43.233413 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerName="ceilometer-notification-agent" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.233422 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerName="ceilometer-notification-agent" Dec 10 19:21:43 crc kubenswrapper[4828]: E1210 19:21:43.233452 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerName="sg-core" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.233461 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerName="sg-core" Dec 10 19:21:43 crc kubenswrapper[4828]: E1210 19:21:43.233479 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerName="proxy-httpd" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.233487 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerName="proxy-httpd" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.233767 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerName="ceilometer-central-agent" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.233813 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerName="sg-core" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.233842 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerName="ceilometer-notification-agent" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.233858 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" containerName="proxy-httpd" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.236544 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.239671 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.240458 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.245512 4828 scope.go:117] "RemoveContainer" containerID="13b30f8eac0d258f61058ba5030d26c003c6447a783aeaf1a92ad703a07b016e" Dec 10 19:21:43 crc kubenswrapper[4828]: E1210 19:21:43.246008 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13b30f8eac0d258f61058ba5030d26c003c6447a783aeaf1a92ad703a07b016e\": container with ID starting with 13b30f8eac0d258f61058ba5030d26c003c6447a783aeaf1a92ad703a07b016e not found: ID does not exist" containerID="13b30f8eac0d258f61058ba5030d26c003c6447a783aeaf1a92ad703a07b016e" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.246037 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13b30f8eac0d258f61058ba5030d26c003c6447a783aeaf1a92ad703a07b016e"} err="failed to get container status \"13b30f8eac0d258f61058ba5030d26c003c6447a783aeaf1a92ad703a07b016e\": rpc error: code = NotFound desc = could not find container \"13b30f8eac0d258f61058ba5030d26c003c6447a783aeaf1a92ad703a07b016e\": container with ID starting with 13b30f8eac0d258f61058ba5030d26c003c6447a783aeaf1a92ad703a07b016e not found: ID does not exist" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.246055 4828 scope.go:117] "RemoveContainer" containerID="8337a232f6d54e5ff0abb55e569be212d7dc79837047c6d1205ea4ee6dbb6686" Dec 10 19:21:43 crc kubenswrapper[4828]: E1210 19:21:43.246305 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8337a232f6d54e5ff0abb55e569be212d7dc79837047c6d1205ea4ee6dbb6686\": container with ID starting with 8337a232f6d54e5ff0abb55e569be212d7dc79837047c6d1205ea4ee6dbb6686 not found: ID does not exist" containerID="8337a232f6d54e5ff0abb55e569be212d7dc79837047c6d1205ea4ee6dbb6686" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.246324 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8337a232f6d54e5ff0abb55e569be212d7dc79837047c6d1205ea4ee6dbb6686"} err="failed to get container status \"8337a232f6d54e5ff0abb55e569be212d7dc79837047c6d1205ea4ee6dbb6686\": rpc error: code = NotFound desc = could not find container \"8337a232f6d54e5ff0abb55e569be212d7dc79837047c6d1205ea4ee6dbb6686\": container with ID starting with 8337a232f6d54e5ff0abb55e569be212d7dc79837047c6d1205ea4ee6dbb6686 not found: ID does not exist" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.246337 4828 scope.go:117] "RemoveContainer" containerID="ee611da9ecabef08e3128b639dd291809c76e57977a77ac75e4bdacd40a37d99" Dec 10 19:21:43 crc kubenswrapper[4828]: E1210 19:21:43.246560 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee611da9ecabef08e3128b639dd291809c76e57977a77ac75e4bdacd40a37d99\": container with ID starting with ee611da9ecabef08e3128b639dd291809c76e57977a77ac75e4bdacd40a37d99 not found: ID does not exist" containerID="ee611da9ecabef08e3128b639dd291809c76e57977a77ac75e4bdacd40a37d99" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.246579 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee611da9ecabef08e3128b639dd291809c76e57977a77ac75e4bdacd40a37d99"} err="failed to get container status \"ee611da9ecabef08e3128b639dd291809c76e57977a77ac75e4bdacd40a37d99\": rpc error: code = NotFound desc = could not find container \"ee611da9ecabef08e3128b639dd291809c76e57977a77ac75e4bdacd40a37d99\": container with ID starting with ee611da9ecabef08e3128b639dd291809c76e57977a77ac75e4bdacd40a37d99 not found: ID does not exist" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.246590 4828 scope.go:117] "RemoveContainer" containerID="bf8d18b5b59b5b77dcb8260d924335e0772a03f59825ce0445c896b38f70f955" Dec 10 19:21:43 crc kubenswrapper[4828]: E1210 19:21:43.247263 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf8d18b5b59b5b77dcb8260d924335e0772a03f59825ce0445c896b38f70f955\": container with ID starting with bf8d18b5b59b5b77dcb8260d924335e0772a03f59825ce0445c896b38f70f955 not found: ID does not exist" containerID="bf8d18b5b59b5b77dcb8260d924335e0772a03f59825ce0445c896b38f70f955" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.247323 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf8d18b5b59b5b77dcb8260d924335e0772a03f59825ce0445c896b38f70f955"} err="failed to get container status \"bf8d18b5b59b5b77dcb8260d924335e0772a03f59825ce0445c896b38f70f955\": rpc error: code = NotFound desc = could not find container \"bf8d18b5b59b5b77dcb8260d924335e0772a03f59825ce0445c896b38f70f955\": container with ID starting with bf8d18b5b59b5b77dcb8260d924335e0772a03f59825ce0445c896b38f70f955 not found: ID does not exist" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.257712 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.420659 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w987w\" (UniqueName: \"kubernetes.io/projected/bef0c776-03a6-4748-a768-0282159d9d89-kube-api-access-w987w\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.420747 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-config-data\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.420782 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.421058 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bef0c776-03a6-4748-a768-0282159d9d89-log-httpd\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.421169 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bef0c776-03a6-4748-a768-0282159d9d89-run-httpd\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.421223 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.421239 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-scripts\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.523260 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w987w\" (UniqueName: \"kubernetes.io/projected/bef0c776-03a6-4748-a768-0282159d9d89-kube-api-access-w987w\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.523330 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-config-data\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.523354 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.523417 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bef0c776-03a6-4748-a768-0282159d9d89-log-httpd\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.523449 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bef0c776-03a6-4748-a768-0282159d9d89-run-httpd\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.523479 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.523492 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-scripts\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.524449 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bef0c776-03a6-4748-a768-0282159d9d89-run-httpd\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.524477 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bef0c776-03a6-4748-a768-0282159d9d89-log-httpd\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.527518 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.529468 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-config-data\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.529545 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.530036 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-scripts\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.547570 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w987w\" (UniqueName: \"kubernetes.io/projected/bef0c776-03a6-4748-a768-0282159d9d89-kube-api-access-w987w\") pod \"ceilometer-0\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.553319 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.791347 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:21:43 crc kubenswrapper[4828]: E1210 19:21:43.792377 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.830680 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f8ce74b-4f95-4e0b-b073-15450cbb4f90" path="/var/lib/kubelet/pods/4f8ce74b-4f95-4e0b-b073-15450cbb4f90/volumes" Dec 10 19:21:43 crc kubenswrapper[4828]: I1210 19:21:43.831628 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f9f6cc8-f9aa-478c-9694-bcc05679ba3a" path="/var/lib/kubelet/pods/5f9f6cc8-f9aa-478c-9694-bcc05679ba3a/volumes" Dec 10 19:21:44 crc kubenswrapper[4828]: W1210 19:21:44.181982 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbef0c776_03a6_4748_a768_0282159d9d89.slice/crio-1a00f8e1b196da2335e1ad38d831c67134ed3b386ca1783b5b6bd6b2d1b65363 WatchSource:0}: Error finding container 1a00f8e1b196da2335e1ad38d831c67134ed3b386ca1783b5b6bd6b2d1b65363: Status 404 returned error can't find the container with id 1a00f8e1b196da2335e1ad38d831c67134ed3b386ca1783b5b6bd6b2d1b65363 Dec 10 19:21:44 crc kubenswrapper[4828]: I1210 19:21:44.185562 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:21:44 crc kubenswrapper[4828]: I1210 19:21:44.851510 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bef0c776-03a6-4748-a768-0282159d9d89","Type":"ContainerStarted","Data":"1a00f8e1b196da2335e1ad38d831c67134ed3b386ca1783b5b6bd6b2d1b65363"} Dec 10 19:21:44 crc kubenswrapper[4828]: I1210 19:21:44.853698 4828 generic.go:334] "Generic (PLEG): container finished" podID="8840b2c3-1f04-4623-8108-eb86d16589cc" containerID="0865111c90502cef9b5f495afd3a3834efc58a21e1ff25f1533ad3ebcced5059" exitCode=0 Dec 10 19:21:44 crc kubenswrapper[4828]: I1210 19:21:44.853752 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6jt6" event={"ID":"8840b2c3-1f04-4623-8108-eb86d16589cc","Type":"ContainerDied","Data":"0865111c90502cef9b5f495afd3a3834efc58a21e1ff25f1533ad3ebcced5059"} Dec 10 19:21:45 crc kubenswrapper[4828]: E1210 19:21:45.343111 4828 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fc35b017bc5e9c9d4621c6f6073a56b6b382e06001a2b8fcd5b8680ffd70a2d5" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 10 19:21:45 crc kubenswrapper[4828]: E1210 19:21:45.348148 4828 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fc35b017bc5e9c9d4621c6f6073a56b6b382e06001a2b8fcd5b8680ffd70a2d5" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 10 19:21:45 crc kubenswrapper[4828]: E1210 19:21:45.351536 4828 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="fc35b017bc5e9c9d4621c6f6073a56b6b382e06001a2b8fcd5b8680ffd70a2d5" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 10 19:21:45 crc kubenswrapper[4828]: E1210 19:21:45.351608 4828 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-787d69597b-vd29s" podUID="abb273e9-4472-4b35-bd92-5c00f7613bd7" containerName="heat-engine" Dec 10 19:21:46 crc kubenswrapper[4828]: I1210 19:21:46.885793 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bef0c776-03a6-4748-a768-0282159d9d89","Type":"ContainerStarted","Data":"dcd38551753ce14e748d492f863e4cf3198593967e87a053eab9d9ed3f7291a0"} Dec 10 19:21:46 crc kubenswrapper[4828]: I1210 19:21:46.886418 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bef0c776-03a6-4748-a768-0282159d9d89","Type":"ContainerStarted","Data":"feb582ff50c2f8993f6aa27e680dfb400895625ac1fc3852f8a7164dfa7b271a"} Dec 10 19:21:46 crc kubenswrapper[4828]: I1210 19:21:46.889445 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6jt6" event={"ID":"8840b2c3-1f04-4623-8108-eb86d16589cc","Type":"ContainerStarted","Data":"64800e9d1d5dbe9bdb95e3ca42c49a13e63e58f9ede4742c5df6e28e83951e74"} Dec 10 19:21:46 crc kubenswrapper[4828]: I1210 19:21:46.913769 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k6jt6" podStartSLOduration=4.582658343 podStartE2EDuration="18.91375344s" podCreationTimestamp="2025-12-10 19:21:28 +0000 UTC" firstStartedPulling="2025-12-10 19:21:31.607098106 +0000 UTC m=+1572.117709101" lastFinishedPulling="2025-12-10 19:21:45.938193193 +0000 UTC m=+1586.448804198" observedRunningTime="2025-12-10 19:21:46.908813019 +0000 UTC m=+1587.419424044" watchObservedRunningTime="2025-12-10 19:21:46.91375344 +0000 UTC m=+1587.424364445" Dec 10 19:21:47 crc kubenswrapper[4828]: I1210 19:21:47.902235 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bef0c776-03a6-4748-a768-0282159d9d89","Type":"ContainerStarted","Data":"422775054cb5b6736a24dbb397a7d66cf27e043e3376efd0c3026b4d4475a918"} Dec 10 19:21:48 crc kubenswrapper[4828]: E1210 19:21:48.256525 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice/crio-cdfddfc4cb8804e6123be5a09adae2421022371cb6d3f95765e08c364825f808\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice\": RecentStats: unable to find data in memory cache]" Dec 10 19:21:48 crc kubenswrapper[4828]: E1210 19:21:48.256774 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice/crio-cdfddfc4cb8804e6123be5a09adae2421022371cb6d3f95765e08c364825f808\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice\": RecentStats: unable to find data in memory cache]" Dec 10 19:21:48 crc kubenswrapper[4828]: I1210 19:21:48.415727 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k6jt6" Dec 10 19:21:48 crc kubenswrapper[4828]: I1210 19:21:48.416385 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k6jt6" Dec 10 19:21:48 crc kubenswrapper[4828]: E1210 19:21:48.904324 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice/crio-cdfddfc4cb8804e6123be5a09adae2421022371cb6d3f95765e08c364825f808\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice\": RecentStats: unable to find data in memory cache]" Dec 10 19:21:48 crc kubenswrapper[4828]: I1210 19:21:48.922941 4828 generic.go:334] "Generic (PLEG): container finished" podID="abb273e9-4472-4b35-bd92-5c00f7613bd7" containerID="fc35b017bc5e9c9d4621c6f6073a56b6b382e06001a2b8fcd5b8680ffd70a2d5" exitCode=0 Dec 10 19:21:48 crc kubenswrapper[4828]: I1210 19:21:48.923028 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-787d69597b-vd29s" event={"ID":"abb273e9-4472-4b35-bd92-5c00f7613bd7","Type":"ContainerDied","Data":"fc35b017bc5e9c9d4621c6f6073a56b6b382e06001a2b8fcd5b8680ffd70a2d5"} Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.337498 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.424818 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-config-data\") pod \"abb273e9-4472-4b35-bd92-5c00f7613bd7\" (UID: \"abb273e9-4472-4b35-bd92-5c00f7613bd7\") " Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.424881 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-config-data-custom\") pod \"abb273e9-4472-4b35-bd92-5c00f7613bd7\" (UID: \"abb273e9-4472-4b35-bd92-5c00f7613bd7\") " Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.425032 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7pk8\" (UniqueName: \"kubernetes.io/projected/abb273e9-4472-4b35-bd92-5c00f7613bd7-kube-api-access-m7pk8\") pod \"abb273e9-4472-4b35-bd92-5c00f7613bd7\" (UID: \"abb273e9-4472-4b35-bd92-5c00f7613bd7\") " Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.425115 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-combined-ca-bundle\") pod \"abb273e9-4472-4b35-bd92-5c00f7613bd7\" (UID: \"abb273e9-4472-4b35-bd92-5c00f7613bd7\") " Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.436055 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abb273e9-4472-4b35-bd92-5c00f7613bd7-kube-api-access-m7pk8" (OuterVolumeSpecName: "kube-api-access-m7pk8") pod "abb273e9-4472-4b35-bd92-5c00f7613bd7" (UID: "abb273e9-4472-4b35-bd92-5c00f7613bd7"). InnerVolumeSpecName "kube-api-access-m7pk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.446067 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "abb273e9-4472-4b35-bd92-5c00f7613bd7" (UID: "abb273e9-4472-4b35-bd92-5c00f7613bd7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.472694 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-k6jt6" podUID="8840b2c3-1f04-4623-8108-eb86d16589cc" containerName="registry-server" probeResult="failure" output=< Dec 10 19:21:49 crc kubenswrapper[4828]: timeout: failed to connect service ":50051" within 1s Dec 10 19:21:49 crc kubenswrapper[4828]: > Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.485012 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "abb273e9-4472-4b35-bd92-5c00f7613bd7" (UID: "abb273e9-4472-4b35-bd92-5c00f7613bd7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.536352 4828 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.536396 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7pk8\" (UniqueName: \"kubernetes.io/projected/abb273e9-4472-4b35-bd92-5c00f7613bd7-kube-api-access-m7pk8\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.536408 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.555953 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-config-data" (OuterVolumeSpecName: "config-data") pod "abb273e9-4472-4b35-bd92-5c00f7613bd7" (UID: "abb273e9-4472-4b35-bd92-5c00f7613bd7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.638005 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abb273e9-4472-4b35-bd92-5c00f7613bd7-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.935067 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-787d69597b-vd29s" event={"ID":"abb273e9-4472-4b35-bd92-5c00f7613bd7","Type":"ContainerDied","Data":"c75a3dd941554e1df1af3c6b9fe656d1458ee4b582e8530c8c25da0e5c18476f"} Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.935138 4828 scope.go:117] "RemoveContainer" containerID="fc35b017bc5e9c9d4621c6f6073a56b6b382e06001a2b8fcd5b8680ffd70a2d5" Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.935104 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-787d69597b-vd29s" Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.938968 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bef0c776-03a6-4748-a768-0282159d9d89","Type":"ContainerStarted","Data":"c7fbb52d4f06c6d2edfde32b416ba59a8864031983ce5be86da03ffccfa27888"} Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.939991 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.966391 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-787d69597b-vd29s"] Dec 10 19:21:49 crc kubenswrapper[4828]: I1210 19:21:49.990869 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-787d69597b-vd29s"] Dec 10 19:21:50 crc kubenswrapper[4828]: I1210 19:21:50.000238 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.570936886 podStartE2EDuration="7.000219613s" podCreationTimestamp="2025-12-10 19:21:43 +0000 UTC" firstStartedPulling="2025-12-10 19:21:44.18458591 +0000 UTC m=+1584.695196915" lastFinishedPulling="2025-12-10 19:21:48.613868637 +0000 UTC m=+1589.124479642" observedRunningTime="2025-12-10 19:21:49.983200522 +0000 UTC m=+1590.493811527" watchObservedRunningTime="2025-12-10 19:21:50.000219613 +0000 UTC m=+1590.510830618" Dec 10 19:21:51 crc kubenswrapper[4828]: I1210 19:21:51.818639 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abb273e9-4472-4b35-bd92-5c00f7613bd7" path="/var/lib/kubelet/pods/abb273e9-4472-4b35-bd92-5c00f7613bd7/volumes" Dec 10 19:21:55 crc kubenswrapper[4828]: I1210 19:21:54.999908 4828 generic.go:334] "Generic (PLEG): container finished" podID="ff038e8c-07a0-41ca-9719-3b8e326721a0" containerID="c0129558dfdd3e4177e968047ac2960485d6ab7fb4131ec5025c1a85d544ca18" exitCode=0 Dec 10 19:21:55 crc kubenswrapper[4828]: I1210 19:21:54.999981 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mzd68" event={"ID":"ff038e8c-07a0-41ca-9719-3b8e326721a0","Type":"ContainerDied","Data":"c0129558dfdd3e4177e968047ac2960485d6ab7fb4131ec5025c1a85d544ca18"} Dec 10 19:21:56 crc kubenswrapper[4828]: I1210 19:21:56.542111 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mzd68" Dec 10 19:21:56 crc kubenswrapper[4828]: I1210 19:21:56.595130 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-config-data\") pod \"ff038e8c-07a0-41ca-9719-3b8e326721a0\" (UID: \"ff038e8c-07a0-41ca-9719-3b8e326721a0\") " Dec 10 19:21:56 crc kubenswrapper[4828]: I1210 19:21:56.595232 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9d95\" (UniqueName: \"kubernetes.io/projected/ff038e8c-07a0-41ca-9719-3b8e326721a0-kube-api-access-j9d95\") pod \"ff038e8c-07a0-41ca-9719-3b8e326721a0\" (UID: \"ff038e8c-07a0-41ca-9719-3b8e326721a0\") " Dec 10 19:21:56 crc kubenswrapper[4828]: I1210 19:21:56.595377 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-scripts\") pod \"ff038e8c-07a0-41ca-9719-3b8e326721a0\" (UID: \"ff038e8c-07a0-41ca-9719-3b8e326721a0\") " Dec 10 19:21:56 crc kubenswrapper[4828]: I1210 19:21:56.595521 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-combined-ca-bundle\") pod \"ff038e8c-07a0-41ca-9719-3b8e326721a0\" (UID: \"ff038e8c-07a0-41ca-9719-3b8e326721a0\") " Dec 10 19:21:56 crc kubenswrapper[4828]: I1210 19:21:56.601663 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-scripts" (OuterVolumeSpecName: "scripts") pod "ff038e8c-07a0-41ca-9719-3b8e326721a0" (UID: "ff038e8c-07a0-41ca-9719-3b8e326721a0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:56 crc kubenswrapper[4828]: I1210 19:21:56.603716 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff038e8c-07a0-41ca-9719-3b8e326721a0-kube-api-access-j9d95" (OuterVolumeSpecName: "kube-api-access-j9d95") pod "ff038e8c-07a0-41ca-9719-3b8e326721a0" (UID: "ff038e8c-07a0-41ca-9719-3b8e326721a0"). InnerVolumeSpecName "kube-api-access-j9d95". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:21:56 crc kubenswrapper[4828]: I1210 19:21:56.631952 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff038e8c-07a0-41ca-9719-3b8e326721a0" (UID: "ff038e8c-07a0-41ca-9719-3b8e326721a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:56 crc kubenswrapper[4828]: I1210 19:21:56.631934 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-config-data" (OuterVolumeSpecName: "config-data") pod "ff038e8c-07a0-41ca-9719-3b8e326721a0" (UID: "ff038e8c-07a0-41ca-9719-3b8e326721a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:21:56 crc kubenswrapper[4828]: I1210 19:21:56.702581 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:56 crc kubenswrapper[4828]: I1210 19:21:56.702828 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9d95\" (UniqueName: \"kubernetes.io/projected/ff038e8c-07a0-41ca-9719-3b8e326721a0-kube-api-access-j9d95\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:56 crc kubenswrapper[4828]: I1210 19:21:56.702968 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:56 crc kubenswrapper[4828]: I1210 19:21:56.703132 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff038e8c-07a0-41ca-9719-3b8e326721a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:21:56 crc kubenswrapper[4828]: I1210 19:21:56.788342 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:21:56 crc kubenswrapper[4828]: E1210 19:21:56.788933 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:21:56 crc kubenswrapper[4828]: E1210 19:21:56.914566 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice/crio-cdfddfc4cb8804e6123be5a09adae2421022371cb6d3f95765e08c364825f808\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice\": RecentStats: unable to find data in memory cache]" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.032342 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-mzd68" event={"ID":"ff038e8c-07a0-41ca-9719-3b8e326721a0","Type":"ContainerDied","Data":"90a30594231e75e9ca3bdbd4f0801e306e052122be51be16baa3c1c3f41508da"} Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.032388 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90a30594231e75e9ca3bdbd4f0801e306e052122be51be16baa3c1c3f41508da" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.032410 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-mzd68" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.137829 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 19:21:57 crc kubenswrapper[4828]: E1210 19:21:57.138541 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff038e8c-07a0-41ca-9719-3b8e326721a0" containerName="nova-cell0-conductor-db-sync" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.138566 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff038e8c-07a0-41ca-9719-3b8e326721a0" containerName="nova-cell0-conductor-db-sync" Dec 10 19:21:57 crc kubenswrapper[4828]: E1210 19:21:57.138628 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abb273e9-4472-4b35-bd92-5c00f7613bd7" containerName="heat-engine" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.138636 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="abb273e9-4472-4b35-bd92-5c00f7613bd7" containerName="heat-engine" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.138917 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="abb273e9-4472-4b35-bd92-5c00f7613bd7" containerName="heat-engine" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.138953 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff038e8c-07a0-41ca-9719-3b8e326721a0" containerName="nova-cell0-conductor-db-sync" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.140636 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.143158 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-j4mkf" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.143360 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.148459 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.213256 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk4qt\" (UniqueName: \"kubernetes.io/projected/860f3307-f863-4231-bf61-e71c70821e2a-kube-api-access-kk4qt\") pod \"nova-cell0-conductor-0\" (UID: \"860f3307-f863-4231-bf61-e71c70821e2a\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.213310 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/860f3307-f863-4231-bf61-e71c70821e2a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"860f3307-f863-4231-bf61-e71c70821e2a\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.213393 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/860f3307-f863-4231-bf61-e71c70821e2a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"860f3307-f863-4231-bf61-e71c70821e2a\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.315369 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk4qt\" (UniqueName: \"kubernetes.io/projected/860f3307-f863-4231-bf61-e71c70821e2a-kube-api-access-kk4qt\") pod \"nova-cell0-conductor-0\" (UID: \"860f3307-f863-4231-bf61-e71c70821e2a\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.315459 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/860f3307-f863-4231-bf61-e71c70821e2a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"860f3307-f863-4231-bf61-e71c70821e2a\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.315726 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/860f3307-f863-4231-bf61-e71c70821e2a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"860f3307-f863-4231-bf61-e71c70821e2a\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.320158 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/860f3307-f863-4231-bf61-e71c70821e2a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"860f3307-f863-4231-bf61-e71c70821e2a\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.320679 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/860f3307-f863-4231-bf61-e71c70821e2a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"860f3307-f863-4231-bf61-e71c70821e2a\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.332277 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk4qt\" (UniqueName: \"kubernetes.io/projected/860f3307-f863-4231-bf61-e71c70821e2a-kube-api-access-kk4qt\") pod \"nova-cell0-conductor-0\" (UID: \"860f3307-f863-4231-bf61-e71c70821e2a\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:21:57 crc kubenswrapper[4828]: I1210 19:21:57.466078 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 10 19:21:58 crc kubenswrapper[4828]: I1210 19:21:58.464300 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k6jt6" Dec 10 19:21:58 crc kubenswrapper[4828]: I1210 19:21:58.512403 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k6jt6" Dec 10 19:21:58 crc kubenswrapper[4828]: E1210 19:21:58.965231 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice/crio-cdfddfc4cb8804e6123be5a09adae2421022371cb6d3f95765e08c364825f808\": RecentStats: unable to find data in memory cache]" Dec 10 19:21:59 crc kubenswrapper[4828]: I1210 19:21:59.316370 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k6jt6"] Dec 10 19:22:00 crc kubenswrapper[4828]: I1210 19:22:00.060524 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k6jt6" podUID="8840b2c3-1f04-4623-8108-eb86d16589cc" containerName="registry-server" containerID="cri-o://64800e9d1d5dbe9bdb95e3ca42c49a13e63e58f9ede4742c5df6e28e83951e74" gracePeriod=2 Dec 10 19:22:00 crc kubenswrapper[4828]: W1210 19:22:00.522964 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod860f3307_f863_4231_bf61_e71c70821e2a.slice/crio-a18f9a16b214b14825782f44f2ae2d5d7efae446ecd05a7836c90992ba42d972 WatchSource:0}: Error finding container a18f9a16b214b14825782f44f2ae2d5d7efae446ecd05a7836c90992ba42d972: Status 404 returned error can't find the container with id a18f9a16b214b14825782f44f2ae2d5d7efae446ecd05a7836c90992ba42d972 Dec 10 19:22:00 crc kubenswrapper[4828]: I1210 19:22:00.530820 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.034228 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k6jt6" Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.071771 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"860f3307-f863-4231-bf61-e71c70821e2a","Type":"ContainerStarted","Data":"a18f9a16b214b14825782f44f2ae2d5d7efae446ecd05a7836c90992ba42d972"} Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.074712 4828 generic.go:334] "Generic (PLEG): container finished" podID="8840b2c3-1f04-4623-8108-eb86d16589cc" containerID="64800e9d1d5dbe9bdb95e3ca42c49a13e63e58f9ede4742c5df6e28e83951e74" exitCode=0 Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.074763 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6jt6" event={"ID":"8840b2c3-1f04-4623-8108-eb86d16589cc","Type":"ContainerDied","Data":"64800e9d1d5dbe9bdb95e3ca42c49a13e63e58f9ede4742c5df6e28e83951e74"} Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.074812 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k6jt6" event={"ID":"8840b2c3-1f04-4623-8108-eb86d16589cc","Type":"ContainerDied","Data":"bd21b7a12881ce8d8c4b33bf4974ded2fb3583fa1ac6fe17e5864f9cd8270974"} Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.074835 4828 scope.go:117] "RemoveContainer" containerID="64800e9d1d5dbe9bdb95e3ca42c49a13e63e58f9ede4742c5df6e28e83951e74" Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.074837 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k6jt6" Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.096844 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zmsn\" (UniqueName: \"kubernetes.io/projected/8840b2c3-1f04-4623-8108-eb86d16589cc-kube-api-access-8zmsn\") pod \"8840b2c3-1f04-4623-8108-eb86d16589cc\" (UID: \"8840b2c3-1f04-4623-8108-eb86d16589cc\") " Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.097001 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8840b2c3-1f04-4623-8108-eb86d16589cc-catalog-content\") pod \"8840b2c3-1f04-4623-8108-eb86d16589cc\" (UID: \"8840b2c3-1f04-4623-8108-eb86d16589cc\") " Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.097040 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8840b2c3-1f04-4623-8108-eb86d16589cc-utilities\") pod \"8840b2c3-1f04-4623-8108-eb86d16589cc\" (UID: \"8840b2c3-1f04-4623-8108-eb86d16589cc\") " Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.098179 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8840b2c3-1f04-4623-8108-eb86d16589cc-utilities" (OuterVolumeSpecName: "utilities") pod "8840b2c3-1f04-4623-8108-eb86d16589cc" (UID: "8840b2c3-1f04-4623-8108-eb86d16589cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.099320 4828 scope.go:117] "RemoveContainer" containerID="0865111c90502cef9b5f495afd3a3834efc58a21e1ff25f1533ad3ebcced5059" Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.102126 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8840b2c3-1f04-4623-8108-eb86d16589cc-kube-api-access-8zmsn" (OuterVolumeSpecName: "kube-api-access-8zmsn") pod "8840b2c3-1f04-4623-8108-eb86d16589cc" (UID: "8840b2c3-1f04-4623-8108-eb86d16589cc"). InnerVolumeSpecName "kube-api-access-8zmsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.122027 4828 scope.go:117] "RemoveContainer" containerID="7637cb5284a5795603af03d3334931232414b132a7807dc97a95d16ece9e8e8a" Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.145496 4828 scope.go:117] "RemoveContainer" containerID="64800e9d1d5dbe9bdb95e3ca42c49a13e63e58f9ede4742c5df6e28e83951e74" Dec 10 19:22:01 crc kubenswrapper[4828]: E1210 19:22:01.145938 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64800e9d1d5dbe9bdb95e3ca42c49a13e63e58f9ede4742c5df6e28e83951e74\": container with ID starting with 64800e9d1d5dbe9bdb95e3ca42c49a13e63e58f9ede4742c5df6e28e83951e74 not found: ID does not exist" containerID="64800e9d1d5dbe9bdb95e3ca42c49a13e63e58f9ede4742c5df6e28e83951e74" Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.145974 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64800e9d1d5dbe9bdb95e3ca42c49a13e63e58f9ede4742c5df6e28e83951e74"} err="failed to get container status \"64800e9d1d5dbe9bdb95e3ca42c49a13e63e58f9ede4742c5df6e28e83951e74\": rpc error: code = NotFound desc = could not find container \"64800e9d1d5dbe9bdb95e3ca42c49a13e63e58f9ede4742c5df6e28e83951e74\": container with ID starting with 64800e9d1d5dbe9bdb95e3ca42c49a13e63e58f9ede4742c5df6e28e83951e74 not found: ID does not exist" Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.145993 4828 scope.go:117] "RemoveContainer" containerID="0865111c90502cef9b5f495afd3a3834efc58a21e1ff25f1533ad3ebcced5059" Dec 10 19:22:01 crc kubenswrapper[4828]: E1210 19:22:01.146994 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0865111c90502cef9b5f495afd3a3834efc58a21e1ff25f1533ad3ebcced5059\": container with ID starting with 0865111c90502cef9b5f495afd3a3834efc58a21e1ff25f1533ad3ebcced5059 not found: ID does not exist" containerID="0865111c90502cef9b5f495afd3a3834efc58a21e1ff25f1533ad3ebcced5059" Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.147053 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0865111c90502cef9b5f495afd3a3834efc58a21e1ff25f1533ad3ebcced5059"} err="failed to get container status \"0865111c90502cef9b5f495afd3a3834efc58a21e1ff25f1533ad3ebcced5059\": rpc error: code = NotFound desc = could not find container \"0865111c90502cef9b5f495afd3a3834efc58a21e1ff25f1533ad3ebcced5059\": container with ID starting with 0865111c90502cef9b5f495afd3a3834efc58a21e1ff25f1533ad3ebcced5059 not found: ID does not exist" Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.147088 4828 scope.go:117] "RemoveContainer" containerID="7637cb5284a5795603af03d3334931232414b132a7807dc97a95d16ece9e8e8a" Dec 10 19:22:01 crc kubenswrapper[4828]: E1210 19:22:01.147420 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7637cb5284a5795603af03d3334931232414b132a7807dc97a95d16ece9e8e8a\": container with ID starting with 7637cb5284a5795603af03d3334931232414b132a7807dc97a95d16ece9e8e8a not found: ID does not exist" containerID="7637cb5284a5795603af03d3334931232414b132a7807dc97a95d16ece9e8e8a" Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.147446 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7637cb5284a5795603af03d3334931232414b132a7807dc97a95d16ece9e8e8a"} err="failed to get container status \"7637cb5284a5795603af03d3334931232414b132a7807dc97a95d16ece9e8e8a\": rpc error: code = NotFound desc = could not find container \"7637cb5284a5795603af03d3334931232414b132a7807dc97a95d16ece9e8e8a\": container with ID starting with 7637cb5284a5795603af03d3334931232414b132a7807dc97a95d16ece9e8e8a not found: ID does not exist" Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.152613 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8840b2c3-1f04-4623-8108-eb86d16589cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8840b2c3-1f04-4623-8108-eb86d16589cc" (UID: "8840b2c3-1f04-4623-8108-eb86d16589cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.200261 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8840b2c3-1f04-4623-8108-eb86d16589cc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.200335 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8840b2c3-1f04-4623-8108-eb86d16589cc-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.200346 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zmsn\" (UniqueName: \"kubernetes.io/projected/8840b2c3-1f04-4623-8108-eb86d16589cc-kube-api-access-8zmsn\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.420029 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k6jt6"] Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.430987 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k6jt6"] Dec 10 19:22:01 crc kubenswrapper[4828]: I1210 19:22:01.802564 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8840b2c3-1f04-4623-8108-eb86d16589cc" path="/var/lib/kubelet/pods/8840b2c3-1f04-4623-8108-eb86d16589cc/volumes" Dec 10 19:22:02 crc kubenswrapper[4828]: I1210 19:22:02.087227 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"860f3307-f863-4231-bf61-e71c70821e2a","Type":"ContainerStarted","Data":"b25269503a6063695d2f58c6405d12d812e495ce0a08ab209d320516e49f14bf"} Dec 10 19:22:02 crc kubenswrapper[4828]: I1210 19:22:02.087377 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 10 19:22:02 crc kubenswrapper[4828]: I1210 19:22:02.119000 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=5.118973218 podStartE2EDuration="5.118973218s" podCreationTimestamp="2025-12-10 19:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:22:02.108178343 +0000 UTC m=+1602.618789348" watchObservedRunningTime="2025-12-10 19:22:02.118973218 +0000 UTC m=+1602.629584223" Dec 10 19:22:03 crc kubenswrapper[4828]: I1210 19:22:03.834348 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:22:03 crc kubenswrapper[4828]: I1210 19:22:03.835372 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bef0c776-03a6-4748-a768-0282159d9d89" containerName="ceilometer-central-agent" containerID="cri-o://feb582ff50c2f8993f6aa27e680dfb400895625ac1fc3852f8a7164dfa7b271a" gracePeriod=30 Dec 10 19:22:03 crc kubenswrapper[4828]: I1210 19:22:03.835554 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bef0c776-03a6-4748-a768-0282159d9d89" containerName="ceilometer-notification-agent" containerID="cri-o://dcd38551753ce14e748d492f863e4cf3198593967e87a053eab9d9ed3f7291a0" gracePeriod=30 Dec 10 19:22:03 crc kubenswrapper[4828]: I1210 19:22:03.835618 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bef0c776-03a6-4748-a768-0282159d9d89" containerName="sg-core" containerID="cri-o://422775054cb5b6736a24dbb397a7d66cf27e043e3376efd0c3026b4d4475a918" gracePeriod=30 Dec 10 19:22:03 crc kubenswrapper[4828]: I1210 19:22:03.835839 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bef0c776-03a6-4748-a768-0282159d9d89" containerName="proxy-httpd" containerID="cri-o://c7fbb52d4f06c6d2edfde32b416ba59a8864031983ce5be86da03ffccfa27888" gracePeriod=30 Dec 10 19:22:03 crc kubenswrapper[4828]: I1210 19:22:03.847100 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="bef0c776-03a6-4748-a768-0282159d9d89" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.228:3000/\": EOF" Dec 10 19:22:04 crc kubenswrapper[4828]: I1210 19:22:04.114355 4828 generic.go:334] "Generic (PLEG): container finished" podID="bef0c776-03a6-4748-a768-0282159d9d89" containerID="c7fbb52d4f06c6d2edfde32b416ba59a8864031983ce5be86da03ffccfa27888" exitCode=0 Dec 10 19:22:04 crc kubenswrapper[4828]: I1210 19:22:04.114672 4828 generic.go:334] "Generic (PLEG): container finished" podID="bef0c776-03a6-4748-a768-0282159d9d89" containerID="422775054cb5b6736a24dbb397a7d66cf27e043e3376efd0c3026b4d4475a918" exitCode=2 Dec 10 19:22:04 crc kubenswrapper[4828]: I1210 19:22:04.114438 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bef0c776-03a6-4748-a768-0282159d9d89","Type":"ContainerDied","Data":"c7fbb52d4f06c6d2edfde32b416ba59a8864031983ce5be86da03ffccfa27888"} Dec 10 19:22:04 crc kubenswrapper[4828]: I1210 19:22:04.114710 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bef0c776-03a6-4748-a768-0282159d9d89","Type":"ContainerDied","Data":"422775054cb5b6736a24dbb397a7d66cf27e043e3376efd0c3026b4d4475a918"} Dec 10 19:22:05 crc kubenswrapper[4828]: I1210 19:22:05.127762 4828 generic.go:334] "Generic (PLEG): container finished" podID="bef0c776-03a6-4748-a768-0282159d9d89" containerID="feb582ff50c2f8993f6aa27e680dfb400895625ac1fc3852f8a7164dfa7b271a" exitCode=0 Dec 10 19:22:05 crc kubenswrapper[4828]: I1210 19:22:05.127848 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bef0c776-03a6-4748-a768-0282159d9d89","Type":"ContainerDied","Data":"feb582ff50c2f8993f6aa27e680dfb400895625ac1fc3852f8a7164dfa7b271a"} Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.016598 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.017204 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="860f3307-f863-4231-bf61-e71c70821e2a" containerName="nova-cell0-conductor-conductor" containerID="cri-o://b25269503a6063695d2f58c6405d12d812e495ce0a08ab209d320516e49f14bf" gracePeriod=30 Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.049603 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.140072 4828 generic.go:334] "Generic (PLEG): container finished" podID="bef0c776-03a6-4748-a768-0282159d9d89" containerID="dcd38551753ce14e748d492f863e4cf3198593967e87a053eab9d9ed3f7291a0" exitCode=0 Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.140125 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bef0c776-03a6-4748-a768-0282159d9d89","Type":"ContainerDied","Data":"dcd38551753ce14e748d492f863e4cf3198593967e87a053eab9d9ed3f7291a0"} Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.797067 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-qkkq9"] Dec 10 19:22:06 crc kubenswrapper[4828]: E1210 19:22:06.797837 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8840b2c3-1f04-4623-8108-eb86d16589cc" containerName="extract-content" Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.797850 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="8840b2c3-1f04-4623-8108-eb86d16589cc" containerName="extract-content" Dec 10 19:22:06 crc kubenswrapper[4828]: E1210 19:22:06.797858 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8840b2c3-1f04-4623-8108-eb86d16589cc" containerName="registry-server" Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.797865 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="8840b2c3-1f04-4623-8108-eb86d16589cc" containerName="registry-server" Dec 10 19:22:06 crc kubenswrapper[4828]: E1210 19:22:06.797882 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8840b2c3-1f04-4623-8108-eb86d16589cc" containerName="extract-utilities" Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.797888 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="8840b2c3-1f04-4623-8108-eb86d16589cc" containerName="extract-utilities" Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.798144 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="8840b2c3-1f04-4623-8108-eb86d16589cc" containerName="registry-server" Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.800517 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qkkq9" Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.806905 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.807143 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.817630 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qkkq9"] Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.956359 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf5xk\" (UniqueName: \"kubernetes.io/projected/a5e369bc-b3a1-444c-a16d-e70c0b754cef-kube-api-access-xf5xk\") pod \"nova-cell0-cell-mapping-qkkq9\" (UID: \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\") " pod="openstack/nova-cell0-cell-mapping-qkkq9" Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.956527 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-scripts\") pod \"nova-cell0-cell-mapping-qkkq9\" (UID: \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\") " pod="openstack/nova-cell0-cell-mapping-qkkq9" Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.956607 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-config-data\") pod \"nova-cell0-cell-mapping-qkkq9\" (UID: \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\") " pod="openstack/nova-cell0-cell-mapping-qkkq9" Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.956627 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qkkq9\" (UID: \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\") " pod="openstack/nova-cell0-cell-mapping-qkkq9" Dec 10 19:22:06 crc kubenswrapper[4828]: I1210 19:22:06.980657 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.048896 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 19:22:07 crc kubenswrapper[4828]: E1210 19:22:07.049417 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bef0c776-03a6-4748-a768-0282159d9d89" containerName="ceilometer-notification-agent" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.049434 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="bef0c776-03a6-4748-a768-0282159d9d89" containerName="ceilometer-notification-agent" Dec 10 19:22:07 crc kubenswrapper[4828]: E1210 19:22:07.049449 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bef0c776-03a6-4748-a768-0282159d9d89" containerName="proxy-httpd" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.049456 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="bef0c776-03a6-4748-a768-0282159d9d89" containerName="proxy-httpd" Dec 10 19:22:07 crc kubenswrapper[4828]: E1210 19:22:07.049477 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bef0c776-03a6-4748-a768-0282159d9d89" containerName="sg-core" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.049483 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="bef0c776-03a6-4748-a768-0282159d9d89" containerName="sg-core" Dec 10 19:22:07 crc kubenswrapper[4828]: E1210 19:22:07.049498 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bef0c776-03a6-4748-a768-0282159d9d89" containerName="ceilometer-central-agent" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.049504 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="bef0c776-03a6-4748-a768-0282159d9d89" containerName="ceilometer-central-agent" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.049711 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="bef0c776-03a6-4748-a768-0282159d9d89" containerName="ceilometer-notification-agent" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.049733 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="bef0c776-03a6-4748-a768-0282159d9d89" containerName="ceilometer-central-agent" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.049745 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="bef0c776-03a6-4748-a768-0282159d9d89" containerName="sg-core" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.049759 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="bef0c776-03a6-4748-a768-0282159d9d89" containerName="proxy-httpd" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.051002 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.056148 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.058561 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf5xk\" (UniqueName: \"kubernetes.io/projected/a5e369bc-b3a1-444c-a16d-e70c0b754cef-kube-api-access-xf5xk\") pod \"nova-cell0-cell-mapping-qkkq9\" (UID: \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\") " pod="openstack/nova-cell0-cell-mapping-qkkq9" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.058646 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-scripts\") pod \"nova-cell0-cell-mapping-qkkq9\" (UID: \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\") " pod="openstack/nova-cell0-cell-mapping-qkkq9" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.058735 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-config-data\") pod \"nova-cell0-cell-mapping-qkkq9\" (UID: \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\") " pod="openstack/nova-cell0-cell-mapping-qkkq9" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.058759 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qkkq9\" (UID: \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\") " pod="openstack/nova-cell0-cell-mapping-qkkq9" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.071901 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qkkq9\" (UID: \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\") " pod="openstack/nova-cell0-cell-mapping-qkkq9" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.074964 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-scripts\") pod \"nova-cell0-cell-mapping-qkkq9\" (UID: \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\") " pod="openstack/nova-cell0-cell-mapping-qkkq9" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.084430 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-config-data\") pod \"nova-cell0-cell-mapping-qkkq9\" (UID: \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\") " pod="openstack/nova-cell0-cell-mapping-qkkq9" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.093564 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.095189 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.105066 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.124217 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-j8658"] Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.125747 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-j8658" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.144309 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf5xk\" (UniqueName: \"kubernetes.io/projected/a5e369bc-b3a1-444c-a16d-e70c0b754cef-kube-api-access-xf5xk\") pod \"nova-cell0-cell-mapping-qkkq9\" (UID: \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\") " pod="openstack/nova-cell0-cell-mapping-qkkq9" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.170730 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-config-data\") pod \"bef0c776-03a6-4748-a768-0282159d9d89\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.170775 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-scripts\") pod \"bef0c776-03a6-4748-a768-0282159d9d89\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.170826 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w987w\" (UniqueName: \"kubernetes.io/projected/bef0c776-03a6-4748-a768-0282159d9d89-kube-api-access-w987w\") pod \"bef0c776-03a6-4748-a768-0282159d9d89\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.170842 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-sg-core-conf-yaml\") pod \"bef0c776-03a6-4748-a768-0282159d9d89\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.171004 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bef0c776-03a6-4748-a768-0282159d9d89-log-httpd\") pod \"bef0c776-03a6-4748-a768-0282159d9d89\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.171081 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bef0c776-03a6-4748-a768-0282159d9d89-run-httpd\") pod \"bef0c776-03a6-4748-a768-0282159d9d89\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.171159 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-combined-ca-bundle\") pod \"bef0c776-03a6-4748-a768-0282159d9d89\" (UID: \"bef0c776-03a6-4748-a768-0282159d9d89\") " Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.171658 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/366eb842-07b0-456f-95c2-d5f93556def8-config-data\") pod \"nova-api-0\" (UID: \"366eb842-07b0-456f-95c2-d5f93556def8\") " pod="openstack/nova-api-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.171932 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjxqt\" (UniqueName: \"kubernetes.io/projected/366eb842-07b0-456f-95c2-d5f93556def8-kube-api-access-hjxqt\") pod \"nova-api-0\" (UID: \"366eb842-07b0-456f-95c2-d5f93556def8\") " pod="openstack/nova-api-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.171959 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366eb842-07b0-456f-95c2-d5f93556def8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"366eb842-07b0-456f-95c2-d5f93556def8\") " pod="openstack/nova-api-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.171993 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/366eb842-07b0-456f-95c2-d5f93556def8-logs\") pod \"nova-api-0\" (UID: \"366eb842-07b0-456f-95c2-d5f93556def8\") " pod="openstack/nova-api-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.173173 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bef0c776-03a6-4748-a768-0282159d9d89-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bef0c776-03a6-4748-a768-0282159d9d89" (UID: "bef0c776-03a6-4748-a768-0282159d9d89"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.189909 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-scripts" (OuterVolumeSpecName: "scripts") pod "bef0c776-03a6-4748-a768-0282159d9d89" (UID: "bef0c776-03a6-4748-a768-0282159d9d89"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.211690 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bef0c776-03a6-4748-a768-0282159d9d89-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bef0c776-03a6-4748-a768-0282159d9d89" (UID: "bef0c776-03a6-4748-a768-0282159d9d89"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.226466 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bef0c776-03a6-4748-a768-0282159d9d89-kube-api-access-w987w" (OuterVolumeSpecName: "kube-api-access-w987w") pod "bef0c776-03a6-4748-a768-0282159d9d89" (UID: "bef0c776-03a6-4748-a768-0282159d9d89"). InnerVolumeSpecName "kube-api-access-w987w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.250898 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.251563 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qkkq9" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.263931 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bef0c776-03a6-4748-a768-0282159d9d89" (UID: "bef0c776-03a6-4748-a768-0282159d9d89"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.288753 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh24x\" (UniqueName: \"kubernetes.io/projected/86fca083-3b11-4d59-9a0c-8099d8257622-kube-api-access-gh24x\") pod \"nova-scheduler-0\" (UID: \"86fca083-3b11-4d59-9a0c-8099d8257622\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.288904 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9a81ca2-b1ab-4f73-a928-332933d6dd23-operator-scripts\") pod \"aodh-db-create-j8658\" (UID: \"b9a81ca2-b1ab-4f73-a928-332933d6dd23\") " pod="openstack/aodh-db-create-j8658" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.289078 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjxqt\" (UniqueName: \"kubernetes.io/projected/366eb842-07b0-456f-95c2-d5f93556def8-kube-api-access-hjxqt\") pod \"nova-api-0\" (UID: \"366eb842-07b0-456f-95c2-d5f93556def8\") " pod="openstack/nova-api-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.289105 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86fca083-3b11-4d59-9a0c-8099d8257622-config-data\") pod \"nova-scheduler-0\" (UID: \"86fca083-3b11-4d59-9a0c-8099d8257622\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.289153 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366eb842-07b0-456f-95c2-d5f93556def8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"366eb842-07b0-456f-95c2-d5f93556def8\") " pod="openstack/nova-api-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.289197 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/366eb842-07b0-456f-95c2-d5f93556def8-logs\") pod \"nova-api-0\" (UID: \"366eb842-07b0-456f-95c2-d5f93556def8\") " pod="openstack/nova-api-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.289312 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzg4j\" (UniqueName: \"kubernetes.io/projected/b9a81ca2-b1ab-4f73-a928-332933d6dd23-kube-api-access-gzg4j\") pod \"aodh-db-create-j8658\" (UID: \"b9a81ca2-b1ab-4f73-a928-332933d6dd23\") " pod="openstack/aodh-db-create-j8658" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.289344 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/366eb842-07b0-456f-95c2-d5f93556def8-config-data\") pod \"nova-api-0\" (UID: \"366eb842-07b0-456f-95c2-d5f93556def8\") " pod="openstack/nova-api-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.289390 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86fca083-3b11-4d59-9a0c-8099d8257622-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"86fca083-3b11-4d59-9a0c-8099d8257622\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.289462 4828 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bef0c776-03a6-4748-a768-0282159d9d89-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.289473 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.289482 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w987w\" (UniqueName: \"kubernetes.io/projected/bef0c776-03a6-4748-a768-0282159d9d89-kube-api-access-w987w\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.289492 4828 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.289501 4828 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bef0c776-03a6-4748-a768-0282159d9d89-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.290042 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/366eb842-07b0-456f-95c2-d5f93556def8-logs\") pod \"nova-api-0\" (UID: \"366eb842-07b0-456f-95c2-d5f93556def8\") " pod="openstack/nova-api-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.295230 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bef0c776-03a6-4748-a768-0282159d9d89","Type":"ContainerDied","Data":"1a00f8e1b196da2335e1ad38d831c67134ed3b386ca1783b5b6bd6b2d1b65363"} Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.295292 4828 scope.go:117] "RemoveContainer" containerID="c7fbb52d4f06c6d2edfde32b416ba59a8864031983ce5be86da03ffccfa27888" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.295447 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.317413 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/366eb842-07b0-456f-95c2-d5f93556def8-config-data\") pod \"nova-api-0\" (UID: \"366eb842-07b0-456f-95c2-d5f93556def8\") " pod="openstack/nova-api-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.318446 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366eb842-07b0-456f-95c2-d5f93556def8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"366eb842-07b0-456f-95c2-d5f93556def8\") " pod="openstack/nova-api-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.338196 4828 generic.go:334] "Generic (PLEG): container finished" podID="860f3307-f863-4231-bf61-e71c70821e2a" containerID="b25269503a6063695d2f58c6405d12d812e495ce0a08ab209d320516e49f14bf" exitCode=0 Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.338354 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"860f3307-f863-4231-bf61-e71c70821e2a","Type":"ContainerDied","Data":"b25269503a6063695d2f58c6405d12d812e495ce0a08ab209d320516e49f14bf"} Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.339549 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"860f3307-f863-4231-bf61-e71c70821e2a","Type":"ContainerDied","Data":"a18f9a16b214b14825782f44f2ae2d5d7efae446ecd05a7836c90992ba42d972"} Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.339594 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a18f9a16b214b14825782f44f2ae2d5d7efae446ecd05a7836c90992ba42d972" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.366167 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.387111 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjxqt\" (UniqueName: \"kubernetes.io/projected/366eb842-07b0-456f-95c2-d5f93556def8-kube-api-access-hjxqt\") pod \"nova-api-0\" (UID: \"366eb842-07b0-456f-95c2-d5f93556def8\") " pod="openstack/nova-api-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.393022 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86fca083-3b11-4d59-9a0c-8099d8257622-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"86fca083-3b11-4d59-9a0c-8099d8257622\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.393305 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh24x\" (UniqueName: \"kubernetes.io/projected/86fca083-3b11-4d59-9a0c-8099d8257622-kube-api-access-gh24x\") pod \"nova-scheduler-0\" (UID: \"86fca083-3b11-4d59-9a0c-8099d8257622\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.393445 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9a81ca2-b1ab-4f73-a928-332933d6dd23-operator-scripts\") pod \"aodh-db-create-j8658\" (UID: \"b9a81ca2-b1ab-4f73-a928-332933d6dd23\") " pod="openstack/aodh-db-create-j8658" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.393647 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86fca083-3b11-4d59-9a0c-8099d8257622-config-data\") pod \"nova-scheduler-0\" (UID: \"86fca083-3b11-4d59-9a0c-8099d8257622\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.393901 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzg4j\" (UniqueName: \"kubernetes.io/projected/b9a81ca2-b1ab-4f73-a928-332933d6dd23-kube-api-access-gzg4j\") pod \"aodh-db-create-j8658\" (UID: \"b9a81ca2-b1ab-4f73-a928-332933d6dd23\") " pod="openstack/aodh-db-create-j8658" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.404927 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9a81ca2-b1ab-4f73-a928-332933d6dd23-operator-scripts\") pod \"aodh-db-create-j8658\" (UID: \"b9a81ca2-b1ab-4f73-a928-332933d6dd23\") " pod="openstack/aodh-db-create-j8658" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.421302 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86fca083-3b11-4d59-9a0c-8099d8257622-config-data\") pod \"nova-scheduler-0\" (UID: \"86fca083-3b11-4d59-9a0c-8099d8257622\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.432262 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86fca083-3b11-4d59-9a0c-8099d8257622-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"86fca083-3b11-4d59-9a0c-8099d8257622\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.463879 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzg4j\" (UniqueName: \"kubernetes.io/projected/b9a81ca2-b1ab-4f73-a928-332933d6dd23-kube-api-access-gzg4j\") pod \"aodh-db-create-j8658\" (UID: \"b9a81ca2-b1ab-4f73-a928-332933d6dd23\") " pod="openstack/aodh-db-create-j8658" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.466528 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh24x\" (UniqueName: \"kubernetes.io/projected/86fca083-3b11-4d59-9a0c-8099d8257622-kube-api-access-gh24x\") pod \"nova-scheduler-0\" (UID: \"86fca083-3b11-4d59-9a0c-8099d8257622\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.490961 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-j8658"] Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.500309 4828 scope.go:117] "RemoveContainer" containerID="422775054cb5b6736a24dbb397a7d66cf27e043e3376efd0c3026b4d4475a918" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.504203 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kk4qt\" (UniqueName: \"kubernetes.io/projected/860f3307-f863-4231-bf61-e71c70821e2a-kube-api-access-kk4qt\") pod \"860f3307-f863-4231-bf61-e71c70821e2a\" (UID: \"860f3307-f863-4231-bf61-e71c70821e2a\") " Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.505755 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/860f3307-f863-4231-bf61-e71c70821e2a-combined-ca-bundle\") pod \"860f3307-f863-4231-bf61-e71c70821e2a\" (UID: \"860f3307-f863-4231-bf61-e71c70821e2a\") " Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.506423 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/860f3307-f863-4231-bf61-e71c70821e2a-config-data\") pod \"860f3307-f863-4231-bf61-e71c70821e2a\" (UID: \"860f3307-f863-4231-bf61-e71c70821e2a\") " Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.509237 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-config-data" (OuterVolumeSpecName: "config-data") pod "bef0c776-03a6-4748-a768-0282159d9d89" (UID: "bef0c776-03a6-4748-a768-0282159d9d89"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.531641 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.554382 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.561008 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/860f3307-f863-4231-bf61-e71c70821e2a-kube-api-access-kk4qt" (OuterVolumeSpecName: "kube-api-access-kk4qt") pod "860f3307-f863-4231-bf61-e71c70821e2a" (UID: "860f3307-f863-4231-bf61-e71c70821e2a"). InnerVolumeSpecName "kube-api-access-kk4qt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.575443 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:22:07 crc kubenswrapper[4828]: E1210 19:22:07.576029 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="860f3307-f863-4231-bf61-e71c70821e2a" containerName="nova-cell0-conductor-conductor" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.576047 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="860f3307-f863-4231-bf61-e71c70821e2a" containerName="nova-cell0-conductor-conductor" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.576337 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="860f3307-f863-4231-bf61-e71c70821e2a" containerName="nova-cell0-conductor-conductor" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.605454 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.611403 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.611438 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kk4qt\" (UniqueName: \"kubernetes.io/projected/860f3307-f863-4231-bf61-e71c70821e2a-kube-api-access-kk4qt\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.612173 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.622935 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.640597 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.653142 4828 scope.go:117] "RemoveContainer" containerID="dcd38551753ce14e748d492f863e4cf3198593967e87a053eab9d9ed3f7291a0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.690894 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/860f3307-f863-4231-bf61-e71c70821e2a-config-data" (OuterVolumeSpecName: "config-data") pod "860f3307-f863-4231-bf61-e71c70821e2a" (UID: "860f3307-f863-4231-bf61-e71c70821e2a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.690996 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/860f3307-f863-4231-bf61-e71c70821e2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "860f3307-f863-4231-bf61-e71c70821e2a" (UID: "860f3307-f863-4231-bf61-e71c70821e2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.703016 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-j8658" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.714011 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07160fa2-a003-4436-a117-b79d5f5627c4-logs\") pod \"nova-metadata-0\" (UID: \"07160fa2-a003-4436-a117-b79d5f5627c4\") " pod="openstack/nova-metadata-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.720164 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07160fa2-a003-4436-a117-b79d5f5627c4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"07160fa2-a003-4436-a117-b79d5f5627c4\") " pod="openstack/nova-metadata-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.720276 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2r8j\" (UniqueName: \"kubernetes.io/projected/07160fa2-a003-4436-a117-b79d5f5627c4-kube-api-access-d2r8j\") pod \"nova-metadata-0\" (UID: \"07160fa2-a003-4436-a117-b79d5f5627c4\") " pod="openstack/nova-metadata-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.721064 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07160fa2-a003-4436-a117-b79d5f5627c4-config-data\") pod \"nova-metadata-0\" (UID: \"07160fa2-a003-4436-a117-b79d5f5627c4\") " pod="openstack/nova-metadata-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.721192 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/860f3307-f863-4231-bf61-e71c70821e2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.721233 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/860f3307-f863-4231-bf61-e71c70821e2a-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.731517 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bef0c776-03a6-4748-a768-0282159d9d89" (UID: "bef0c776-03a6-4748-a768-0282159d9d89"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.757019 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-be42-account-create-update-j8bfv"] Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.767064 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-be42-account-create-update-j8bfv" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.772087 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.780993 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-be42-account-create-update-j8bfv"] Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.823047 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07160fa2-a003-4436-a117-b79d5f5627c4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"07160fa2-a003-4436-a117-b79d5f5627c4\") " pod="openstack/nova-metadata-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.823495 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2r8j\" (UniqueName: \"kubernetes.io/projected/07160fa2-a003-4436-a117-b79d5f5627c4-kube-api-access-d2r8j\") pod \"nova-metadata-0\" (UID: \"07160fa2-a003-4436-a117-b79d5f5627c4\") " pod="openstack/nova-metadata-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.823585 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07160fa2-a003-4436-a117-b79d5f5627c4-config-data\") pod \"nova-metadata-0\" (UID: \"07160fa2-a003-4436-a117-b79d5f5627c4\") " pod="openstack/nova-metadata-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.823634 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07160fa2-a003-4436-a117-b79d5f5627c4-logs\") pod \"nova-metadata-0\" (UID: \"07160fa2-a003-4436-a117-b79d5f5627c4\") " pod="openstack/nova-metadata-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.823787 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bef0c776-03a6-4748-a768-0282159d9d89-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.824505 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07160fa2-a003-4436-a117-b79d5f5627c4-logs\") pod \"nova-metadata-0\" (UID: \"07160fa2-a003-4436-a117-b79d5f5627c4\") " pod="openstack/nova-metadata-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.838964 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07160fa2-a003-4436-a117-b79d5f5627c4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"07160fa2-a003-4436-a117-b79d5f5627c4\") " pod="openstack/nova-metadata-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.849597 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07160fa2-a003-4436-a117-b79d5f5627c4-config-data\") pod \"nova-metadata-0\" (UID: \"07160fa2-a003-4436-a117-b79d5f5627c4\") " pod="openstack/nova-metadata-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.858908 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2r8j\" (UniqueName: \"kubernetes.io/projected/07160fa2-a003-4436-a117-b79d5f5627c4-kube-api-access-d2r8j\") pod \"nova-metadata-0\" (UID: \"07160fa2-a003-4436-a117-b79d5f5627c4\") " pod="openstack/nova-metadata-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.864485 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-ncsqd"] Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.866974 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.877079 4828 scope.go:117] "RemoveContainer" containerID="feb582ff50c2f8993f6aa27e680dfb400895625ac1fc3852f8a7164dfa7b271a" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.896120 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.897921 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.944543 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptg6v\" (UniqueName: \"kubernetes.io/projected/cf3c61d9-4edd-4fe5-a591-7b3ef8455b13-kube-api-access-ptg6v\") pod \"aodh-be42-account-create-update-j8bfv\" (UID: \"cf3c61d9-4edd-4fe5-a591-7b3ef8455b13\") " pod="openstack/aodh-be42-account-create-update-j8bfv" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.944703 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf3c61d9-4edd-4fe5-a591-7b3ef8455b13-operator-scripts\") pod \"aodh-be42-account-create-update-j8bfv\" (UID: \"cf3c61d9-4edd-4fe5-a591-7b3ef8455b13\") " pod="openstack/aodh-be42-account-create-update-j8bfv" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.948842 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-ncsqd"] Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.951708 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.979989 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:22:07 crc kubenswrapper[4828]: I1210 19:22:07.991542 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.054732 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-config\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.054788 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqs72\" (UniqueName: \"kubernetes.io/projected/78cfbb25-b948-41f2-9157-97b4f9a70ecf-kube-api-access-rqs72\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.054866 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf3c61d9-4edd-4fe5-a591-7b3ef8455b13-operator-scripts\") pod \"aodh-be42-account-create-update-j8bfv\" (UID: \"cf3c61d9-4edd-4fe5-a591-7b3ef8455b13\") " pod="openstack/aodh-be42-account-create-update-j8bfv" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.054899 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8crbq\" (UniqueName: \"kubernetes.io/projected/529a0e7b-83de-4c8f-aeb6-4575868cdf74-kube-api-access-8crbq\") pod \"nova-cell1-novncproxy-0\" (UID: \"529a0e7b-83de-4c8f-aeb6-4575868cdf74\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.054924 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.054976 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.055008 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.055304 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/529a0e7b-83de-4c8f-aeb6-4575868cdf74-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"529a0e7b-83de-4c8f-aeb6-4575868cdf74\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.055336 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/529a0e7b-83de-4c8f-aeb6-4575868cdf74-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"529a0e7b-83de-4c8f-aeb6-4575868cdf74\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.055380 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.055412 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptg6v\" (UniqueName: \"kubernetes.io/projected/cf3c61d9-4edd-4fe5-a591-7b3ef8455b13-kube-api-access-ptg6v\") pod \"aodh-be42-account-create-update-j8bfv\" (UID: \"cf3c61d9-4edd-4fe5-a591-7b3ef8455b13\") " pod="openstack/aodh-be42-account-create-update-j8bfv" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.059504 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf3c61d9-4edd-4fe5-a591-7b3ef8455b13-operator-scripts\") pod \"aodh-be42-account-create-update-j8bfv\" (UID: \"cf3c61d9-4edd-4fe5-a591-7b3ef8455b13\") " pod="openstack/aodh-be42-account-create-update-j8bfv" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.100282 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptg6v\" (UniqueName: \"kubernetes.io/projected/cf3c61d9-4edd-4fe5-a591-7b3ef8455b13-kube-api-access-ptg6v\") pod \"aodh-be42-account-create-update-j8bfv\" (UID: \"cf3c61d9-4edd-4fe5-a591-7b3ef8455b13\") " pod="openstack/aodh-be42-account-create-update-j8bfv" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.118834 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.132681 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.143736 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.148102 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.151347 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.151431 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.159189 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.160443 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.160527 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-config\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.160547 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqs72\" (UniqueName: \"kubernetes.io/projected/78cfbb25-b948-41f2-9157-97b4f9a70ecf-kube-api-access-rqs72\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.160594 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8crbq\" (UniqueName: \"kubernetes.io/projected/529a0e7b-83de-4c8f-aeb6-4575868cdf74-kube-api-access-8crbq\") pod \"nova-cell1-novncproxy-0\" (UID: \"529a0e7b-83de-4c8f-aeb6-4575868cdf74\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.160613 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.160644 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.160667 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.160939 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/529a0e7b-83de-4c8f-aeb6-4575868cdf74-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"529a0e7b-83de-4c8f-aeb6-4575868cdf74\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.160964 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/529a0e7b-83de-4c8f-aeb6-4575868cdf74-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"529a0e7b-83de-4c8f-aeb6-4575868cdf74\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.162212 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.162471 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.162673 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.163067 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-config\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.163284 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.169617 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/529a0e7b-83de-4c8f-aeb6-4575868cdf74-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"529a0e7b-83de-4c8f-aeb6-4575868cdf74\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.197758 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8crbq\" (UniqueName: \"kubernetes.io/projected/529a0e7b-83de-4c8f-aeb6-4575868cdf74-kube-api-access-8crbq\") pod \"nova-cell1-novncproxy-0\" (UID: \"529a0e7b-83de-4c8f-aeb6-4575868cdf74\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.206310 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/529a0e7b-83de-4c8f-aeb6-4575868cdf74-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"529a0e7b-83de-4c8f-aeb6-4575868cdf74\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.209181 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqs72\" (UniqueName: \"kubernetes.io/projected/78cfbb25-b948-41f2-9157-97b4f9a70ecf-kube-api-access-rqs72\") pod \"dnsmasq-dns-568d7fd7cf-ncsqd\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.269385 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8f736d0-0707-4e40-af46-4e6882b62586-run-httpd\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.269454 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmfmd\" (UniqueName: \"kubernetes.io/projected/a8f736d0-0707-4e40-af46-4e6882b62586-kube-api-access-jmfmd\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.269548 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.269664 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.269743 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-scripts\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.269762 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8f736d0-0707-4e40-af46-4e6882b62586-log-httpd\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.270035 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-config-data\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.286287 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-be42-account-create-update-j8bfv" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.298910 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.348978 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qkkq9"] Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.362680 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.370595 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.376515 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-scripts\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.376573 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8f736d0-0707-4e40-af46-4e6882b62586-log-httpd\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.376669 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-config-data\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.376760 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8f736d0-0707-4e40-af46-4e6882b62586-run-httpd\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.376790 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmfmd\" (UniqueName: \"kubernetes.io/projected/a8f736d0-0707-4e40-af46-4e6882b62586-kube-api-access-jmfmd\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.376996 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.377146 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.377535 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8f736d0-0707-4e40-af46-4e6882b62586-log-httpd\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.378227 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8f736d0-0707-4e40-af46-4e6882b62586-run-httpd\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.395576 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.397087 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-config-data\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.398011 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.407854 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wr6ll"] Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.410296 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-scripts\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.411166 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmfmd\" (UniqueName: \"kubernetes.io/projected/a8f736d0-0707-4e40-af46-4e6882b62586-kube-api-access-jmfmd\") pod \"ceilometer-0\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.415213 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-wr6ll" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.429744 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.430114 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.446013 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wr6ll"] Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.491632 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.506016 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.518282 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.541403 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.542974 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.547067 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.564194 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.588280 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-scripts\") pod \"nova-cell1-conductor-db-sync-wr6ll\" (UID: \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\") " pod="openstack/nova-cell1-conductor-db-sync-wr6ll" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.588391 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89t5p\" (UniqueName: \"kubernetes.io/projected/f0793df8-d37f-4645-8f1d-3cc501d1ce40-kube-api-access-89t5p\") pod \"nova-cell1-conductor-db-sync-wr6ll\" (UID: \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\") " pod="openstack/nova-cell1-conductor-db-sync-wr6ll" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.588477 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-config-data\") pod \"nova-cell1-conductor-db-sync-wr6ll\" (UID: \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\") " pod="openstack/nova-cell1-conductor-db-sync-wr6ll" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.588510 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-wr6ll\" (UID: \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\") " pod="openstack/nova-cell1-conductor-db-sync-wr6ll" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.690859 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v2pg\" (UniqueName: \"kubernetes.io/projected/5c3fe039-ce8f-4d89-8357-727c36514bf8-kube-api-access-6v2pg\") pod \"nova-cell0-conductor-0\" (UID: \"5c3fe039-ce8f-4d89-8357-727c36514bf8\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.691309 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-config-data\") pod \"nova-cell1-conductor-db-sync-wr6ll\" (UID: \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\") " pod="openstack/nova-cell1-conductor-db-sync-wr6ll" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.691378 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-wr6ll\" (UID: \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\") " pod="openstack/nova-cell1-conductor-db-sync-wr6ll" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.691417 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c3fe039-ce8f-4d89-8357-727c36514bf8-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5c3fe039-ce8f-4d89-8357-727c36514bf8\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.691570 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-scripts\") pod \"nova-cell1-conductor-db-sync-wr6ll\" (UID: \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\") " pod="openstack/nova-cell1-conductor-db-sync-wr6ll" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.691623 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c3fe039-ce8f-4d89-8357-727c36514bf8-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5c3fe039-ce8f-4d89-8357-727c36514bf8\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.691686 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89t5p\" (UniqueName: \"kubernetes.io/projected/f0793df8-d37f-4645-8f1d-3cc501d1ce40-kube-api-access-89t5p\") pod \"nova-cell1-conductor-db-sync-wr6ll\" (UID: \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\") " pod="openstack/nova-cell1-conductor-db-sync-wr6ll" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.700082 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-config-data\") pod \"nova-cell1-conductor-db-sync-wr6ll\" (UID: \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\") " pod="openstack/nova-cell1-conductor-db-sync-wr6ll" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.703413 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-scripts\") pod \"nova-cell1-conductor-db-sync-wr6ll\" (UID: \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\") " pod="openstack/nova-cell1-conductor-db-sync-wr6ll" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.704140 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-wr6ll\" (UID: \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\") " pod="openstack/nova-cell1-conductor-db-sync-wr6ll" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.724143 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89t5p\" (UniqueName: \"kubernetes.io/projected/f0793df8-d37f-4645-8f1d-3cc501d1ce40-kube-api-access-89t5p\") pod \"nova-cell1-conductor-db-sync-wr6ll\" (UID: \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\") " pod="openstack/nova-cell1-conductor-db-sync-wr6ll" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.742550 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.780012 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.792942 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v2pg\" (UniqueName: \"kubernetes.io/projected/5c3fe039-ce8f-4d89-8357-727c36514bf8-kube-api-access-6v2pg\") pod \"nova-cell0-conductor-0\" (UID: \"5c3fe039-ce8f-4d89-8357-727c36514bf8\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.793011 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c3fe039-ce8f-4d89-8357-727c36514bf8-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5c3fe039-ce8f-4d89-8357-727c36514bf8\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.793108 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c3fe039-ce8f-4d89-8357-727c36514bf8-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5c3fe039-ce8f-4d89-8357-727c36514bf8\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.797502 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-wr6ll" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.797897 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c3fe039-ce8f-4d89-8357-727c36514bf8-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5c3fe039-ce8f-4d89-8357-727c36514bf8\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.801382 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c3fe039-ce8f-4d89-8357-727c36514bf8-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5c3fe039-ce8f-4d89-8357-727c36514bf8\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.803016 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-j8658"] Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.827052 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v2pg\" (UniqueName: \"kubernetes.io/projected/5c3fe039-ce8f-4d89-8357-727c36514bf8-kube-api-access-6v2pg\") pod \"nova-cell0-conductor-0\" (UID: \"5c3fe039-ce8f-4d89-8357-727c36514bf8\") " pod="openstack/nova-cell0-conductor-0" Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.886600 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:22:08 crc kubenswrapper[4828]: I1210 19:22:08.986515 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:09.199681 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:09.235565 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-ncsqd"] Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:09.375932 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-be42-account-create-update-j8bfv"] Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:09.394042 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-j8658" event={"ID":"b9a81ca2-b1ab-4f73-a928-332933d6dd23","Type":"ContainerStarted","Data":"5c4d6477387947edb0c421a01054cdf917c21c3a38f7a0d8a05bdcc45d8ef1b1"} Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:09.394103 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-j8658" event={"ID":"b9a81ca2-b1ab-4f73-a928-332933d6dd23","Type":"ContainerStarted","Data":"ad9823869d5987ddffd8afebf93fbebafddc20078b58334ad04273a0441121ed"} Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:09.415648 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qkkq9" event={"ID":"a5e369bc-b3a1-444c-a16d-e70c0b754cef","Type":"ContainerStarted","Data":"4cba3f00e1c0289ab6e191fb3c7d9b20adfe56571354534de26f2d9cd52096a4"} Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:09.415701 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qkkq9" event={"ID":"a5e369bc-b3a1-444c-a16d-e70c0b754cef","Type":"ContainerStarted","Data":"a6cb039ea4accc6311a83a9d67c310829e5af256b76d97abfc136794d62a5224"} Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:09.432575 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"529a0e7b-83de-4c8f-aeb6-4575868cdf74","Type":"ContainerStarted","Data":"fc36113aaf6c8e162fb91ce4b96a7df30fba0bf5e75d026df9d664c150bf7228"} Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:09.447712 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-create-j8658" podStartSLOduration=3.447635882 podStartE2EDuration="3.447635882s" podCreationTimestamp="2025-12-10 19:22:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:22:09.416678942 +0000 UTC m=+1609.927289947" watchObservedRunningTime="2025-12-10 19:22:09.447635882 +0000 UTC m=+1609.958246887" Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:09.449082 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"366eb842-07b0-456f-95c2-d5f93556def8","Type":"ContainerStarted","Data":"881eca32b1805f190d7ccc823c5561212a90ec51a43905b2ea42eb78ed0a525b"} Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:09.451285 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-qkkq9" podStartSLOduration=3.451270328 podStartE2EDuration="3.451270328s" podCreationTimestamp="2025-12-10 19:22:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:22:09.432965393 +0000 UTC m=+1609.943576398" watchObservedRunningTime="2025-12-10 19:22:09.451270328 +0000 UTC m=+1609.961881333" Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:09.452346 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" event={"ID":"78cfbb25-b948-41f2-9157-97b4f9a70ecf","Type":"ContainerStarted","Data":"4149e3cb3f6009f99de8a8433a94de03f933aa4bfc74c440b6b1b7aeda676461"} Dec 10 19:22:10 crc kubenswrapper[4828]: E1210 19:22:09.453546 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice/crio-cdfddfc4cb8804e6123be5a09adae2421022371cb6d3f95765e08c364825f808\": RecentStats: unable to find data in memory cache]" Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:09.457829 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"86fca083-3b11-4d59-9a0c-8099d8257622","Type":"ContainerStarted","Data":"0a4e1db2900bd5af9a0508698e0dccf59cda9f77465ec37af3c8bdb7015b2ccc"} Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:09.462817 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07160fa2-a003-4436-a117-b79d5f5627c4","Type":"ContainerStarted","Data":"979c0065a7464630a78151441bd02aab96731c1cbbfb54d5bb8ac6f006323dd8"} Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:09.821930 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="860f3307-f863-4231-bf61-e71c70821e2a" path="/var/lib/kubelet/pods/860f3307-f863-4231-bf61-e71c70821e2a/volumes" Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:09.823272 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bef0c776-03a6-4748-a768-0282159d9d89" path="/var/lib/kubelet/pods/bef0c776-03a6-4748-a768-0282159d9d89/volumes" Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:10.359873 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:10.423071 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wr6ll"] Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:10.522107 4828 generic.go:334] "Generic (PLEG): container finished" podID="78cfbb25-b948-41f2-9157-97b4f9a70ecf" containerID="98d08462a5b2fd62e809965335547b0ba03a8c839194ca947320c732a43e7b54" exitCode=0 Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:10.522618 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" event={"ID":"78cfbb25-b948-41f2-9157-97b4f9a70ecf","Type":"ContainerDied","Data":"98d08462a5b2fd62e809965335547b0ba03a8c839194ca947320c732a43e7b54"} Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:10.532424 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-wr6ll" event={"ID":"f0793df8-d37f-4645-8f1d-3cc501d1ce40","Type":"ContainerStarted","Data":"9772372298df53e4a34e99846d688a77d96081b79ea18e0fd7aea91919e460cd"} Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:10.551247 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8f736d0-0707-4e40-af46-4e6882b62586","Type":"ContainerStarted","Data":"2412de5b4686b2fa229ec4beccce1824f81aeca50d9cc7bc1401b2cff980b66a"} Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:10.569739 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:10.570051 4828 generic.go:334] "Generic (PLEG): container finished" podID="b9a81ca2-b1ab-4f73-a928-332933d6dd23" containerID="5c4d6477387947edb0c421a01054cdf917c21c3a38f7a0d8a05bdcc45d8ef1b1" exitCode=0 Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:10.570821 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-j8658" event={"ID":"b9a81ca2-b1ab-4f73-a928-332933d6dd23","Type":"ContainerDied","Data":"5c4d6477387947edb0c421a01054cdf917c21c3a38f7a0d8a05bdcc45d8ef1b1"} Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:10.609602 4828 generic.go:334] "Generic (PLEG): container finished" podID="cf3c61d9-4edd-4fe5-a591-7b3ef8455b13" containerID="878b0bf453660768bb6ad41edd6478b6244747f6bbeca165254f867dcf7a3a1f" exitCode=0 Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:10.611312 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-be42-account-create-update-j8bfv" event={"ID":"cf3c61d9-4edd-4fe5-a591-7b3ef8455b13","Type":"ContainerDied","Data":"878b0bf453660768bb6ad41edd6478b6244747f6bbeca165254f867dcf7a3a1f"} Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:10.611365 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-be42-account-create-update-j8bfv" event={"ID":"cf3c61d9-4edd-4fe5-a591-7b3ef8455b13","Type":"ContainerStarted","Data":"e3cc52ccf9cafa279e3602cd375699c6f7f0efa78930d632ccda11eb96c94f64"} Dec 10 19:22:10 crc kubenswrapper[4828]: I1210 19:22:10.790360 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:22:10 crc kubenswrapper[4828]: E1210 19:22:10.790733 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:22:11 crc kubenswrapper[4828]: I1210 19:22:11.621124 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5c3fe039-ce8f-4d89-8357-727c36514bf8","Type":"ContainerStarted","Data":"d65735cb2d5779bd1581257cb25d8f28c77c80fcc905f49beb40ecc18239c7d3"} Dec 10 19:22:11 crc kubenswrapper[4828]: I1210 19:22:11.625999 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-wr6ll" event={"ID":"f0793df8-d37f-4645-8f1d-3cc501d1ce40","Type":"ContainerStarted","Data":"90281732151b276c44c07e2e044c7066937b465cc73bbef5ac0f509487deaeaf"} Dec 10 19:22:11 crc kubenswrapper[4828]: I1210 19:22:11.648630 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-wr6ll" podStartSLOduration=3.648611952 podStartE2EDuration="3.648611952s" podCreationTimestamp="2025-12-10 19:22:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:22:11.643057295 +0000 UTC m=+1612.153668320" watchObservedRunningTime="2025-12-10 19:22:11.648611952 +0000 UTC m=+1612.159222957" Dec 10 19:22:11 crc kubenswrapper[4828]: E1210 19:22:11.682215 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice/crio-cdfddfc4cb8804e6123be5a09adae2421022371cb6d3f95765e08c364825f808\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice\": RecentStats: unable to find data in memory cache]" Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.486377 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-j8658" Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.493925 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-be42-account-create-update-j8bfv" Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.600002 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptg6v\" (UniqueName: \"kubernetes.io/projected/cf3c61d9-4edd-4fe5-a591-7b3ef8455b13-kube-api-access-ptg6v\") pod \"cf3c61d9-4edd-4fe5-a591-7b3ef8455b13\" (UID: \"cf3c61d9-4edd-4fe5-a591-7b3ef8455b13\") " Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.600165 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzg4j\" (UniqueName: \"kubernetes.io/projected/b9a81ca2-b1ab-4f73-a928-332933d6dd23-kube-api-access-gzg4j\") pod \"b9a81ca2-b1ab-4f73-a928-332933d6dd23\" (UID: \"b9a81ca2-b1ab-4f73-a928-332933d6dd23\") " Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.600224 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9a81ca2-b1ab-4f73-a928-332933d6dd23-operator-scripts\") pod \"b9a81ca2-b1ab-4f73-a928-332933d6dd23\" (UID: \"b9a81ca2-b1ab-4f73-a928-332933d6dd23\") " Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.600406 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf3c61d9-4edd-4fe5-a591-7b3ef8455b13-operator-scripts\") pod \"cf3c61d9-4edd-4fe5-a591-7b3ef8455b13\" (UID: \"cf3c61d9-4edd-4fe5-a591-7b3ef8455b13\") " Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.601371 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf3c61d9-4edd-4fe5-a591-7b3ef8455b13-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cf3c61d9-4edd-4fe5-a591-7b3ef8455b13" (UID: "cf3c61d9-4edd-4fe5-a591-7b3ef8455b13"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.602906 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9a81ca2-b1ab-4f73-a928-332933d6dd23-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b9a81ca2-b1ab-4f73-a928-332933d6dd23" (UID: "b9a81ca2-b1ab-4f73-a928-332933d6dd23"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.606686 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9a81ca2-b1ab-4f73-a928-332933d6dd23-kube-api-access-gzg4j" (OuterVolumeSpecName: "kube-api-access-gzg4j") pod "b9a81ca2-b1ab-4f73-a928-332933d6dd23" (UID: "b9a81ca2-b1ab-4f73-a928-332933d6dd23"). InnerVolumeSpecName "kube-api-access-gzg4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.606981 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf3c61d9-4edd-4fe5-a591-7b3ef8455b13-kube-api-access-ptg6v" (OuterVolumeSpecName: "kube-api-access-ptg6v") pod "cf3c61d9-4edd-4fe5-a591-7b3ef8455b13" (UID: "cf3c61d9-4edd-4fe5-a591-7b3ef8455b13"). InnerVolumeSpecName "kube-api-access-ptg6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.651892 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-j8658" event={"ID":"b9a81ca2-b1ab-4f73-a928-332933d6dd23","Type":"ContainerDied","Data":"ad9823869d5987ddffd8afebf93fbebafddc20078b58334ad04273a0441121ed"} Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.651931 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad9823869d5987ddffd8afebf93fbebafddc20078b58334ad04273a0441121ed" Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.652021 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-j8658" Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.654518 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-be42-account-create-update-j8bfv" event={"ID":"cf3c61d9-4edd-4fe5-a591-7b3ef8455b13","Type":"ContainerDied","Data":"e3cc52ccf9cafa279e3602cd375699c6f7f0efa78930d632ccda11eb96c94f64"} Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.654744 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3cc52ccf9cafa279e3602cd375699c6f7f0efa78930d632ccda11eb96c94f64" Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.654855 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-be42-account-create-update-j8bfv" Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.703262 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzg4j\" (UniqueName: \"kubernetes.io/projected/b9a81ca2-b1ab-4f73-a928-332933d6dd23-kube-api-access-gzg4j\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.703295 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b9a81ca2-b1ab-4f73-a928-332933d6dd23-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.703305 4828 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf3c61d9-4edd-4fe5-a591-7b3ef8455b13-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:13 crc kubenswrapper[4828]: I1210 19:22:13.703313 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptg6v\" (UniqueName: \"kubernetes.io/projected/cf3c61d9-4edd-4fe5-a591-7b3ef8455b13-kube-api-access-ptg6v\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:14 crc kubenswrapper[4828]: I1210 19:22:14.258910 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xlq5k"] Dec 10 19:22:14 crc kubenswrapper[4828]: E1210 19:22:14.259841 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9a81ca2-b1ab-4f73-a928-332933d6dd23" containerName="mariadb-database-create" Dec 10 19:22:14 crc kubenswrapper[4828]: I1210 19:22:14.259870 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9a81ca2-b1ab-4f73-a928-332933d6dd23" containerName="mariadb-database-create" Dec 10 19:22:14 crc kubenswrapper[4828]: E1210 19:22:14.259902 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf3c61d9-4edd-4fe5-a591-7b3ef8455b13" containerName="mariadb-account-create-update" Dec 10 19:22:14 crc kubenswrapper[4828]: I1210 19:22:14.259911 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf3c61d9-4edd-4fe5-a591-7b3ef8455b13" containerName="mariadb-account-create-update" Dec 10 19:22:14 crc kubenswrapper[4828]: I1210 19:22:14.260217 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf3c61d9-4edd-4fe5-a591-7b3ef8455b13" containerName="mariadb-account-create-update" Dec 10 19:22:14 crc kubenswrapper[4828]: I1210 19:22:14.260277 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9a81ca2-b1ab-4f73-a928-332933d6dd23" containerName="mariadb-database-create" Dec 10 19:22:14 crc kubenswrapper[4828]: I1210 19:22:14.262273 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xlq5k" Dec 10 19:22:14 crc kubenswrapper[4828]: I1210 19:22:14.283967 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlq5k"] Dec 10 19:22:14 crc kubenswrapper[4828]: I1210 19:22:14.419436 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a615a974-7cc4-4ed1-957f-9ca4f1024f26-catalog-content\") pod \"redhat-marketplace-xlq5k\" (UID: \"a615a974-7cc4-4ed1-957f-9ca4f1024f26\") " pod="openshift-marketplace/redhat-marketplace-xlq5k" Dec 10 19:22:14 crc kubenswrapper[4828]: I1210 19:22:14.419496 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpgv7\" (UniqueName: \"kubernetes.io/projected/a615a974-7cc4-4ed1-957f-9ca4f1024f26-kube-api-access-qpgv7\") pod \"redhat-marketplace-xlq5k\" (UID: \"a615a974-7cc4-4ed1-957f-9ca4f1024f26\") " pod="openshift-marketplace/redhat-marketplace-xlq5k" Dec 10 19:22:14 crc kubenswrapper[4828]: I1210 19:22:14.419536 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a615a974-7cc4-4ed1-957f-9ca4f1024f26-utilities\") pod \"redhat-marketplace-xlq5k\" (UID: \"a615a974-7cc4-4ed1-957f-9ca4f1024f26\") " pod="openshift-marketplace/redhat-marketplace-xlq5k" Dec 10 19:22:14 crc kubenswrapper[4828]: I1210 19:22:14.522383 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpgv7\" (UniqueName: \"kubernetes.io/projected/a615a974-7cc4-4ed1-957f-9ca4f1024f26-kube-api-access-qpgv7\") pod \"redhat-marketplace-xlq5k\" (UID: \"a615a974-7cc4-4ed1-957f-9ca4f1024f26\") " pod="openshift-marketplace/redhat-marketplace-xlq5k" Dec 10 19:22:14 crc kubenswrapper[4828]: I1210 19:22:14.522477 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a615a974-7cc4-4ed1-957f-9ca4f1024f26-utilities\") pod \"redhat-marketplace-xlq5k\" (UID: \"a615a974-7cc4-4ed1-957f-9ca4f1024f26\") " pod="openshift-marketplace/redhat-marketplace-xlq5k" Dec 10 19:22:14 crc kubenswrapper[4828]: I1210 19:22:14.522762 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a615a974-7cc4-4ed1-957f-9ca4f1024f26-catalog-content\") pod \"redhat-marketplace-xlq5k\" (UID: \"a615a974-7cc4-4ed1-957f-9ca4f1024f26\") " pod="openshift-marketplace/redhat-marketplace-xlq5k" Dec 10 19:22:14 crc kubenswrapper[4828]: I1210 19:22:14.523166 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a615a974-7cc4-4ed1-957f-9ca4f1024f26-utilities\") pod \"redhat-marketplace-xlq5k\" (UID: \"a615a974-7cc4-4ed1-957f-9ca4f1024f26\") " pod="openshift-marketplace/redhat-marketplace-xlq5k" Dec 10 19:22:14 crc kubenswrapper[4828]: I1210 19:22:14.523291 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a615a974-7cc4-4ed1-957f-9ca4f1024f26-catalog-content\") pod \"redhat-marketplace-xlq5k\" (UID: \"a615a974-7cc4-4ed1-957f-9ca4f1024f26\") " pod="openshift-marketplace/redhat-marketplace-xlq5k" Dec 10 19:22:14 crc kubenswrapper[4828]: I1210 19:22:14.546869 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpgv7\" (UniqueName: \"kubernetes.io/projected/a615a974-7cc4-4ed1-957f-9ca4f1024f26-kube-api-access-qpgv7\") pod \"redhat-marketplace-xlq5k\" (UID: \"a615a974-7cc4-4ed1-957f-9ca4f1024f26\") " pod="openshift-marketplace/redhat-marketplace-xlq5k" Dec 10 19:22:14 crc kubenswrapper[4828]: I1210 19:22:14.586157 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xlq5k" Dec 10 19:22:15 crc kubenswrapper[4828]: I1210 19:22:15.682910 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" event={"ID":"78cfbb25-b948-41f2-9157-97b4f9a70ecf","Type":"ContainerStarted","Data":"9b0f0f82dec4572c981ad7cccc4373ae5c762b74bc983a788c415acf024ac396"} Dec 10 19:22:15 crc kubenswrapper[4828]: I1210 19:22:15.683499 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:15 crc kubenswrapper[4828]: I1210 19:22:15.718521 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" podStartSLOduration=8.71849884 podStartE2EDuration="8.71849884s" podCreationTimestamp="2025-12-10 19:22:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:22:15.70527966 +0000 UTC m=+1616.215890665" watchObservedRunningTime="2025-12-10 19:22:15.71849884 +0000 UTC m=+1616.229109865" Dec 10 19:22:15 crc kubenswrapper[4828]: I1210 19:22:15.946240 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlq5k"] Dec 10 19:22:15 crc kubenswrapper[4828]: W1210 19:22:15.984680 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda615a974_7cc4_4ed1_957f_9ca4f1024f26.slice/crio-ad09d920e335be211057ca7e61a231aefabad843081449f3556c190cbdb21e41 WatchSource:0}: Error finding container ad09d920e335be211057ca7e61a231aefabad843081449f3556c190cbdb21e41: Status 404 returned error can't find the container with id ad09d920e335be211057ca7e61a231aefabad843081449f3556c190cbdb21e41 Dec 10 19:22:16 crc kubenswrapper[4828]: I1210 19:22:16.693928 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"86fca083-3b11-4d59-9a0c-8099d8257622","Type":"ContainerStarted","Data":"183616e4449fa53ec1ca869e092100729984d6f74eead5b35b19268e7cef0654"} Dec 10 19:22:16 crc kubenswrapper[4828]: I1210 19:22:16.696665 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8f736d0-0707-4e40-af46-4e6882b62586","Type":"ContainerStarted","Data":"5c6fc4c0290724312e38ce7793e291b58430c0024bfd1582155d38594ec86ecd"} Dec 10 19:22:16 crc kubenswrapper[4828]: I1210 19:22:16.697849 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5c3fe039-ce8f-4d89-8357-727c36514bf8","Type":"ContainerStarted","Data":"f97535514bb663262ad805a857b7769bbd915a902aa3c5008fb0608db0b86d59"} Dec 10 19:22:16 crc kubenswrapper[4828]: I1210 19:22:16.698414 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 10 19:22:16 crc kubenswrapper[4828]: I1210 19:22:16.700258 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07160fa2-a003-4436-a117-b79d5f5627c4","Type":"ContainerStarted","Data":"f4b05950edd0b0b24d7f0423d2bc8320cfbdfcfa3c56d4df9a31fdce6993c881"} Dec 10 19:22:16 crc kubenswrapper[4828]: I1210 19:22:16.700283 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07160fa2-a003-4436-a117-b79d5f5627c4","Type":"ContainerStarted","Data":"e945d8ddd0409a4467a895c081dc91e12683d5639acf76aedd81ce5ddb1c306c"} Dec 10 19:22:16 crc kubenswrapper[4828]: I1210 19:22:16.702056 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"529a0e7b-83de-4c8f-aeb6-4575868cdf74","Type":"ContainerStarted","Data":"b5bcd0007c22828533bdaaeb4721a76296c00808cdee3d5d27358d3075976373"} Dec 10 19:22:16 crc kubenswrapper[4828]: I1210 19:22:16.703640 4828 generic.go:334] "Generic (PLEG): container finished" podID="a615a974-7cc4-4ed1-957f-9ca4f1024f26" containerID="406e8e47de3ab043374488bec2694a7b90ea38d4663ad1b8f2b2ee9539cbac43" exitCode=0 Dec 10 19:22:16 crc kubenswrapper[4828]: I1210 19:22:16.703719 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlq5k" event={"ID":"a615a974-7cc4-4ed1-957f-9ca4f1024f26","Type":"ContainerDied","Data":"406e8e47de3ab043374488bec2694a7b90ea38d4663ad1b8f2b2ee9539cbac43"} Dec 10 19:22:16 crc kubenswrapper[4828]: I1210 19:22:16.703745 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlq5k" event={"ID":"a615a974-7cc4-4ed1-957f-9ca4f1024f26","Type":"ContainerStarted","Data":"ad09d920e335be211057ca7e61a231aefabad843081449f3556c190cbdb21e41"} Dec 10 19:22:16 crc kubenswrapper[4828]: I1210 19:22:16.705450 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"366eb842-07b0-456f-95c2-d5f93556def8","Type":"ContainerStarted","Data":"772a8274bb34ae1d80fcd1523e509e5a334556f577dc97317ed8e41d2d84e955"} Dec 10 19:22:16 crc kubenswrapper[4828]: I1210 19:22:16.705474 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"366eb842-07b0-456f-95c2-d5f93556def8","Type":"ContainerStarted","Data":"240b44e4a825640b5f5b51f3b697b1f15462d0b1d7cbfcdff74f08f7e5863ed7"} Dec 10 19:22:16 crc kubenswrapper[4828]: I1210 19:22:16.725202 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=4.016512809 podStartE2EDuration="10.725178611s" podCreationTimestamp="2025-12-10 19:22:06 +0000 UTC" firstStartedPulling="2025-12-10 19:22:08.785996449 +0000 UTC m=+1609.296607454" lastFinishedPulling="2025-12-10 19:22:15.494662261 +0000 UTC m=+1616.005273256" observedRunningTime="2025-12-10 19:22:16.71570089 +0000 UTC m=+1617.226311915" watchObservedRunningTime="2025-12-10 19:22:16.725178611 +0000 UTC m=+1617.235789616" Dec 10 19:22:16 crc kubenswrapper[4828]: I1210 19:22:16.747505 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.881956394 podStartE2EDuration="10.747487152s" podCreationTimestamp="2025-12-10 19:22:06 +0000 UTC" firstStartedPulling="2025-12-10 19:22:08.770889928 +0000 UTC m=+1609.281500933" lastFinishedPulling="2025-12-10 19:22:15.636420686 +0000 UTC m=+1616.147031691" observedRunningTime="2025-12-10 19:22:16.735166476 +0000 UTC m=+1617.245777481" watchObservedRunningTime="2025-12-10 19:22:16.747487152 +0000 UTC m=+1617.258098157" Dec 10 19:22:16 crc kubenswrapper[4828]: I1210 19:22:16.782968 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=8.782949921 podStartE2EDuration="8.782949921s" podCreationTimestamp="2025-12-10 19:22:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:22:16.776369907 +0000 UTC m=+1617.286980912" watchObservedRunningTime="2025-12-10 19:22:16.782949921 +0000 UTC m=+1617.293560926" Dec 10 19:22:16 crc kubenswrapper[4828]: I1210 19:22:16.794896 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.089717517 podStartE2EDuration="9.794879497s" podCreationTimestamp="2025-12-10 19:22:07 +0000 UTC" firstStartedPulling="2025-12-10 19:22:08.931252566 +0000 UTC m=+1609.441863571" lastFinishedPulling="2025-12-10 19:22:15.636414546 +0000 UTC m=+1616.147025551" observedRunningTime="2025-12-10 19:22:16.790649605 +0000 UTC m=+1617.301260610" watchObservedRunningTime="2025-12-10 19:22:16.794879497 +0000 UTC m=+1617.305490502" Dec 10 19:22:16 crc kubenswrapper[4828]: I1210 19:22:16.817091 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.577490996 podStartE2EDuration="9.817073305s" podCreationTimestamp="2025-12-10 19:22:07 +0000 UTC" firstStartedPulling="2025-12-10 19:22:09.255601696 +0000 UTC m=+1609.766212701" lastFinishedPulling="2025-12-10 19:22:15.495183995 +0000 UTC m=+1616.005795010" observedRunningTime="2025-12-10 19:22:16.811060206 +0000 UTC m=+1617.321671221" watchObservedRunningTime="2025-12-10 19:22:16.817073305 +0000 UTC m=+1617.327684310" Dec 10 19:22:17 crc kubenswrapper[4828]: I1210 19:22:17.555480 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 19:22:17 crc kubenswrapper[4828]: I1210 19:22:17.555522 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 19:22:17 crc kubenswrapper[4828]: I1210 19:22:17.607084 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 10 19:22:17 crc kubenswrapper[4828]: I1210 19:22:17.607748 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 10 19:22:17 crc kubenswrapper[4828]: I1210 19:22:17.638209 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 10 19:22:17 crc kubenswrapper[4828]: I1210 19:22:17.750678 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.373624 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-cell1-novncproxy-0" podUID="529a0e7b-83de-4c8f-aeb6-4575868cdf74" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.0.237:6080/vnc_lite.html\": dial tcp 10.217.0.237:6080: connect: connection refused" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.504237 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.504296 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.504314 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.504354 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.504369 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.504379 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.639069 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="366eb842-07b0-456f-95c2-d5f93556def8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.231:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.639065 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="366eb842-07b0-456f-95c2-d5f93556def8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.231:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.716888 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-dtpxc"] Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.720115 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-dtpxc" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.725705 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.725764 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.726760 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.727184 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-qg8rx" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.749740 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-dtpxc"] Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.843912 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-config-data\") pod \"aodh-db-sync-dtpxc\" (UID: \"057a6200-117a-47be-97df-55ae94e6f763\") " pod="openstack/aodh-db-sync-dtpxc" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.844054 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-combined-ca-bundle\") pod \"aodh-db-sync-dtpxc\" (UID: \"057a6200-117a-47be-97df-55ae94e6f763\") " pod="openstack/aodh-db-sync-dtpxc" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.844084 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-scripts\") pod \"aodh-db-sync-dtpxc\" (UID: \"057a6200-117a-47be-97df-55ae94e6f763\") " pod="openstack/aodh-db-sync-dtpxc" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.844173 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzjlw\" (UniqueName: \"kubernetes.io/projected/057a6200-117a-47be-97df-55ae94e6f763-kube-api-access-vzjlw\") pod \"aodh-db-sync-dtpxc\" (UID: \"057a6200-117a-47be-97df-55ae94e6f763\") " pod="openstack/aodh-db-sync-dtpxc" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.947040 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-config-data\") pod \"aodh-db-sync-dtpxc\" (UID: \"057a6200-117a-47be-97df-55ae94e6f763\") " pod="openstack/aodh-db-sync-dtpxc" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.947466 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-combined-ca-bundle\") pod \"aodh-db-sync-dtpxc\" (UID: \"057a6200-117a-47be-97df-55ae94e6f763\") " pod="openstack/aodh-db-sync-dtpxc" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.947564 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-scripts\") pod \"aodh-db-sync-dtpxc\" (UID: \"057a6200-117a-47be-97df-55ae94e6f763\") " pod="openstack/aodh-db-sync-dtpxc" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.947737 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzjlw\" (UniqueName: \"kubernetes.io/projected/057a6200-117a-47be-97df-55ae94e6f763-kube-api-access-vzjlw\") pod \"aodh-db-sync-dtpxc\" (UID: \"057a6200-117a-47be-97df-55ae94e6f763\") " pod="openstack/aodh-db-sync-dtpxc" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.962765 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-scripts\") pod \"aodh-db-sync-dtpxc\" (UID: \"057a6200-117a-47be-97df-55ae94e6f763\") " pod="openstack/aodh-db-sync-dtpxc" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.964336 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-config-data\") pod \"aodh-db-sync-dtpxc\" (UID: \"057a6200-117a-47be-97df-55ae94e6f763\") " pod="openstack/aodh-db-sync-dtpxc" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.964713 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-combined-ca-bundle\") pod \"aodh-db-sync-dtpxc\" (UID: \"057a6200-117a-47be-97df-55ae94e6f763\") " pod="openstack/aodh-db-sync-dtpxc" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.992958 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="07160fa2-a003-4436-a117-b79d5f5627c4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.0.234:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 19:22:18 crc kubenswrapper[4828]: I1210 19:22:18.993057 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="07160fa2-a003-4436-a117-b79d5f5627c4" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.0.234:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 19:22:19 crc kubenswrapper[4828]: I1210 19:22:19.078070 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzjlw\" (UniqueName: \"kubernetes.io/projected/057a6200-117a-47be-97df-55ae94e6f763-kube-api-access-vzjlw\") pod \"aodh-db-sync-dtpxc\" (UID: \"057a6200-117a-47be-97df-55ae94e6f763\") " pod="openstack/aodh-db-sync-dtpxc" Dec 10 19:22:19 crc kubenswrapper[4828]: I1210 19:22:19.081892 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-dtpxc" Dec 10 19:22:19 crc kubenswrapper[4828]: I1210 19:22:19.645154 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-dtpxc"] Dec 10 19:22:19 crc kubenswrapper[4828]: I1210 19:22:19.645212 4828 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:22:19 crc kubenswrapper[4828]: I1210 19:22:19.756248 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-dtpxc" event={"ID":"057a6200-117a-47be-97df-55ae94e6f763","Type":"ContainerStarted","Data":"64f94a1eca02b850a046b6892a2e8bcf7d297d26a5a11d0cb1ccee60f7b63e19"} Dec 10 19:22:19 crc kubenswrapper[4828]: E1210 19:22:19.811038 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb19ce251_54b7_40f3_bbe0_6afb6064b307.slice/crio-cdfddfc4cb8804e6123be5a09adae2421022371cb6d3f95765e08c364825f808\": RecentStats: unable to find data in memory cache]" Dec 10 19:22:21 crc kubenswrapper[4828]: I1210 19:22:21.783703 4828 generic.go:334] "Generic (PLEG): container finished" podID="a5e369bc-b3a1-444c-a16d-e70c0b754cef" containerID="4cba3f00e1c0289ab6e191fb3c7d9b20adfe56571354534de26f2d9cd52096a4" exitCode=0 Dec 10 19:22:21 crc kubenswrapper[4828]: I1210 19:22:21.783892 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qkkq9" event={"ID":"a5e369bc-b3a1-444c-a16d-e70c0b754cef","Type":"ContainerDied","Data":"4cba3f00e1c0289ab6e191fb3c7d9b20adfe56571354534de26f2d9cd52096a4"} Dec 10 19:22:21 crc kubenswrapper[4828]: I1210 19:22:21.787351 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlq5k" event={"ID":"a615a974-7cc4-4ed1-957f-9ca4f1024f26","Type":"ContainerStarted","Data":"018b75b946941c5269e855f2e35fba21e4f6dc17513a34aa6be23cd06164f41c"} Dec 10 19:22:21 crc kubenswrapper[4828]: I1210 19:22:21.789444 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:22:21 crc kubenswrapper[4828]: E1210 19:22:21.789720 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:22:21 crc kubenswrapper[4828]: I1210 19:22:21.805094 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8f736d0-0707-4e40-af46-4e6882b62586","Type":"ContainerStarted","Data":"8fc37f29001dd68cd353a6955c98a84deb39e1934c5524f98a4993c0aee981ce"} Dec 10 19:22:22 crc kubenswrapper[4828]: I1210 19:22:22.820593 4828 generic.go:334] "Generic (PLEG): container finished" podID="a615a974-7cc4-4ed1-957f-9ca4f1024f26" containerID="018b75b946941c5269e855f2e35fba21e4f6dc17513a34aa6be23cd06164f41c" exitCode=0 Dec 10 19:22:22 crc kubenswrapper[4828]: I1210 19:22:22.820709 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlq5k" event={"ID":"a615a974-7cc4-4ed1-957f-9ca4f1024f26","Type":"ContainerDied","Data":"018b75b946941c5269e855f2e35fba21e4f6dc17513a34aa6be23cd06164f41c"} Dec 10 19:22:22 crc kubenswrapper[4828]: I1210 19:22:22.825981 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8f736d0-0707-4e40-af46-4e6882b62586","Type":"ContainerStarted","Data":"73bb1fae9724fc9bd5012c30e1f40798f43c89cf4eb2c0e820dbd1518eef051c"} Dec 10 19:22:23 crc kubenswrapper[4828]: I1210 19:22:23.300078 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:22:23 crc kubenswrapper[4828]: I1210 19:22:23.371108 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-jlhgg"] Dec 10 19:22:23 crc kubenswrapper[4828]: I1210 19:22:23.371349 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" podUID="41b1aa0c-723c-4f3f-9e52-26328c8ca954" containerName="dnsmasq-dns" containerID="cri-o://532541bc74b35ce5d925fdafc6ed651ece85f11572169103a80f90b238b5d6e2" gracePeriod=10 Dec 10 19:22:23 crc kubenswrapper[4828]: I1210 19:22:23.846908 4828 generic.go:334] "Generic (PLEG): container finished" podID="41b1aa0c-723c-4f3f-9e52-26328c8ca954" containerID="532541bc74b35ce5d925fdafc6ed651ece85f11572169103a80f90b238b5d6e2" exitCode=0 Dec 10 19:22:23 crc kubenswrapper[4828]: I1210 19:22:23.847029 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" event={"ID":"41b1aa0c-723c-4f3f-9e52-26328c8ca954","Type":"ContainerDied","Data":"532541bc74b35ce5d925fdafc6ed651ece85f11572169103a80f90b238b5d6e2"} Dec 10 19:22:24 crc kubenswrapper[4828]: I1210 19:22:24.029460 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 10 19:22:24 crc kubenswrapper[4828]: I1210 19:22:24.560122 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:22:24 crc kubenswrapper[4828]: I1210 19:22:24.561328 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="529a0e7b-83de-4c8f-aeb6-4575868cdf74" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://b5bcd0007c22828533bdaaeb4721a76296c00808cdee3d5d27358d3075976373" gracePeriod=30 Dec 10 19:22:24 crc kubenswrapper[4828]: I1210 19:22:24.582735 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:22:24 crc kubenswrapper[4828]: I1210 19:22:24.583043 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="07160fa2-a003-4436-a117-b79d5f5627c4" containerName="nova-metadata-log" containerID="cri-o://e945d8ddd0409a4467a895c081dc91e12683d5639acf76aedd81ce5ddb1c306c" gracePeriod=30 Dec 10 19:22:24 crc kubenswrapper[4828]: I1210 19:22:24.583361 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="07160fa2-a003-4436-a117-b79d5f5627c4" containerName="nova-metadata-metadata" containerID="cri-o://f4b05950edd0b0b24d7f0423d2bc8320cfbdfcfa3c56d4df9a31fdce6993c881" gracePeriod=30 Dec 10 19:22:24 crc kubenswrapper[4828]: I1210 19:22:24.865118 4828 generic.go:334] "Generic (PLEG): container finished" podID="07160fa2-a003-4436-a117-b79d5f5627c4" containerID="e945d8ddd0409a4467a895c081dc91e12683d5639acf76aedd81ce5ddb1c306c" exitCode=143 Dec 10 19:22:24 crc kubenswrapper[4828]: I1210 19:22:24.865165 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07160fa2-a003-4436-a117-b79d5f5627c4","Type":"ContainerDied","Data":"e945d8ddd0409a4467a895c081dc91e12683d5639acf76aedd81ce5ddb1c306c"} Dec 10 19:22:25 crc kubenswrapper[4828]: I1210 19:22:25.715884 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" podUID="41b1aa0c-723c-4f3f-9e52-26328c8ca954" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.208:5353: connect: connection refused" Dec 10 19:22:25 crc kubenswrapper[4828]: I1210 19:22:25.884297 4828 generic.go:334] "Generic (PLEG): container finished" podID="529a0e7b-83de-4c8f-aeb6-4575868cdf74" containerID="b5bcd0007c22828533bdaaeb4721a76296c00808cdee3d5d27358d3075976373" exitCode=0 Dec 10 19:22:25 crc kubenswrapper[4828]: I1210 19:22:25.884345 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"529a0e7b-83de-4c8f-aeb6-4575868cdf74","Type":"ContainerDied","Data":"b5bcd0007c22828533bdaaeb4721a76296c00808cdee3d5d27358d3075976373"} Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.398379 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qkkq9" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.422036 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.578523 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-scripts\") pod \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\" (UID: \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\") " Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.578996 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-combined-ca-bundle\") pod \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\" (UID: \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\") " Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.579140 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xf5xk\" (UniqueName: \"kubernetes.io/projected/a5e369bc-b3a1-444c-a16d-e70c0b754cef-kube-api-access-xf5xk\") pod \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\" (UID: \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\") " Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.579250 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-config-data\") pod \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\" (UID: \"a5e369bc-b3a1-444c-a16d-e70c0b754cef\") " Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.586897 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5e369bc-b3a1-444c-a16d-e70c0b754cef-kube-api-access-xf5xk" (OuterVolumeSpecName: "kube-api-access-xf5xk") pod "a5e369bc-b3a1-444c-a16d-e70c0b754cef" (UID: "a5e369bc-b3a1-444c-a16d-e70c0b754cef"). InnerVolumeSpecName "kube-api-access-xf5xk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.590962 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-scripts" (OuterVolumeSpecName: "scripts") pod "a5e369bc-b3a1-444c-a16d-e70c0b754cef" (UID: "a5e369bc-b3a1-444c-a16d-e70c0b754cef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.651235 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-config-data" (OuterVolumeSpecName: "config-data") pod "a5e369bc-b3a1-444c-a16d-e70c0b754cef" (UID: "a5e369bc-b3a1-444c-a16d-e70c0b754cef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.668299 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5e369bc-b3a1-444c-a16d-e70c0b754cef" (UID: "a5e369bc-b3a1-444c-a16d-e70c0b754cef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.669609 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.688156 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xf5xk\" (UniqueName: \"kubernetes.io/projected/a5e369bc-b3a1-444c-a16d-e70c0b754cef-kube-api-access-xf5xk\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.688198 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.688209 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.688218 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5e369bc-b3a1-444c-a16d-e70c0b754cef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.782853 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.789077 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8crbq\" (UniqueName: \"kubernetes.io/projected/529a0e7b-83de-4c8f-aeb6-4575868cdf74-kube-api-access-8crbq\") pod \"529a0e7b-83de-4c8f-aeb6-4575868cdf74\" (UID: \"529a0e7b-83de-4c8f-aeb6-4575868cdf74\") " Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.789141 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/529a0e7b-83de-4c8f-aeb6-4575868cdf74-combined-ca-bundle\") pod \"529a0e7b-83de-4c8f-aeb6-4575868cdf74\" (UID: \"529a0e7b-83de-4c8f-aeb6-4575868cdf74\") " Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.789480 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/529a0e7b-83de-4c8f-aeb6-4575868cdf74-config-data\") pod \"529a0e7b-83de-4c8f-aeb6-4575868cdf74\" (UID: \"529a0e7b-83de-4c8f-aeb6-4575868cdf74\") " Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.796560 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/529a0e7b-83de-4c8f-aeb6-4575868cdf74-kube-api-access-8crbq" (OuterVolumeSpecName: "kube-api-access-8crbq") pod "529a0e7b-83de-4c8f-aeb6-4575868cdf74" (UID: "529a0e7b-83de-4c8f-aeb6-4575868cdf74"). InnerVolumeSpecName "kube-api-access-8crbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.838929 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/529a0e7b-83de-4c8f-aeb6-4575868cdf74-config-data" (OuterVolumeSpecName: "config-data") pod "529a0e7b-83de-4c8f-aeb6-4575868cdf74" (UID: "529a0e7b-83de-4c8f-aeb6-4575868cdf74"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.851632 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/529a0e7b-83de-4c8f-aeb6-4575868cdf74-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "529a0e7b-83de-4c8f-aeb6-4575868cdf74" (UID: "529a0e7b-83de-4c8f-aeb6-4575868cdf74"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.893594 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgb7h\" (UniqueName: \"kubernetes.io/projected/41b1aa0c-723c-4f3f-9e52-26328c8ca954-kube-api-access-hgb7h\") pod \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.893672 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-dns-svc\") pod \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.893757 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-ovsdbserver-sb\") pod \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.893811 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-config\") pod \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.893894 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-dns-swift-storage-0\") pod \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.893996 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-ovsdbserver-nb\") pod \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\" (UID: \"41b1aa0c-723c-4f3f-9e52-26328c8ca954\") " Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.902807 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8crbq\" (UniqueName: \"kubernetes.io/projected/529a0e7b-83de-4c8f-aeb6-4575868cdf74-kube-api-access-8crbq\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.902846 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/529a0e7b-83de-4c8f-aeb6-4575868cdf74-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.902861 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/529a0e7b-83de-4c8f-aeb6-4575868cdf74-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.905012 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.905178 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-jlhgg" event={"ID":"41b1aa0c-723c-4f3f-9e52-26328c8ca954","Type":"ContainerDied","Data":"2430290d21788f1db2b4c1e3acde5854f44e131999245d3af1e9b686411297d8"} Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.905740 4828 scope.go:117] "RemoveContainer" containerID="532541bc74b35ce5d925fdafc6ed651ece85f11572169103a80f90b238b5d6e2" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.911353 4828 generic.go:334] "Generic (PLEG): container finished" podID="f0793df8-d37f-4645-8f1d-3cc501d1ce40" containerID="90281732151b276c44c07e2e044c7066937b465cc73bbef5ac0f509487deaeaf" exitCode=0 Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.911454 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-wr6ll" event={"ID":"f0793df8-d37f-4645-8f1d-3cc501d1ce40","Type":"ContainerDied","Data":"90281732151b276c44c07e2e044c7066937b465cc73bbef5ac0f509487deaeaf"} Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.914249 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-dtpxc" event={"ID":"057a6200-117a-47be-97df-55ae94e6f763","Type":"ContainerStarted","Data":"68e5730b588bb1538a77492edf973b42c140868fae4a86ba497bcb09bbca320a"} Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.916668 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"529a0e7b-83de-4c8f-aeb6-4575868cdf74","Type":"ContainerDied","Data":"fc36113aaf6c8e162fb91ce4b96a7df30fba0bf5e75d026df9d664c150bf7228"} Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.916817 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.919328 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qkkq9" event={"ID":"a5e369bc-b3a1-444c-a16d-e70c0b754cef","Type":"ContainerDied","Data":"a6cb039ea4accc6311a83a9d67c310829e5af256b76d97abfc136794d62a5224"} Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.919534 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6cb039ea4accc6311a83a9d67c310829e5af256b76d97abfc136794d62a5224" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.919403 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qkkq9" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.923824 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8f736d0-0707-4e40-af46-4e6882b62586","Type":"ContainerStarted","Data":"5a873c23c32a5dc22e600ce16b6e6ff38985d36e40eb7913b342b61e9887f0ee"} Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.924159 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.947059 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41b1aa0c-723c-4f3f-9e52-26328c8ca954-kube-api-access-hgb7h" (OuterVolumeSpecName: "kube-api-access-hgb7h") pod "41b1aa0c-723c-4f3f-9e52-26328c8ca954" (UID: "41b1aa0c-723c-4f3f-9e52-26328c8ca954"). InnerVolumeSpecName "kube-api-access-hgb7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.971234 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-dtpxc" podStartSLOduration=2.196676831 podStartE2EDuration="8.971210429s" podCreationTimestamp="2025-12-10 19:22:18 +0000 UTC" firstStartedPulling="2025-12-10 19:22:19.644919318 +0000 UTC m=+1620.155530323" lastFinishedPulling="2025-12-10 19:22:26.419452916 +0000 UTC m=+1626.930063921" observedRunningTime="2025-12-10 19:22:26.948557729 +0000 UTC m=+1627.459168734" watchObservedRunningTime="2025-12-10 19:22:26.971210429 +0000 UTC m=+1627.481821434" Dec 10 19:22:26 crc kubenswrapper[4828]: I1210 19:22:26.975451 4828 scope.go:117] "RemoveContainer" containerID="9be33178bb0a13451b3f987b140e3b14a6d716ce4cc174a6a394d6c6af0ef3bc" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.005725 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgb7h\" (UniqueName: \"kubernetes.io/projected/41b1aa0c-723c-4f3f-9e52-26328c8ca954-kube-api-access-hgb7h\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.006708 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.9876151760000003 podStartE2EDuration="19.006694388s" podCreationTimestamp="2025-12-10 19:22:08 +0000 UTC" firstStartedPulling="2025-12-10 19:22:10.398264488 +0000 UTC m=+1610.908875493" lastFinishedPulling="2025-12-10 19:22:26.4173437 +0000 UTC m=+1626.927954705" observedRunningTime="2025-12-10 19:22:26.965063206 +0000 UTC m=+1627.475674211" watchObservedRunningTime="2025-12-10 19:22:27.006694388 +0000 UTC m=+1627.517305393" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.025973 4828 scope.go:117] "RemoveContainer" containerID="b5bcd0007c22828533bdaaeb4721a76296c00808cdee3d5d27358d3075976373" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.059269 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.087165 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.092401 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-config" (OuterVolumeSpecName: "config") pod "41b1aa0c-723c-4f3f-9e52-26328c8ca954" (UID: "41b1aa0c-723c-4f3f-9e52-26328c8ca954"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.094758 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "41b1aa0c-723c-4f3f-9e52-26328c8ca954" (UID: "41b1aa0c-723c-4f3f-9e52-26328c8ca954"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.096333 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "41b1aa0c-723c-4f3f-9e52-26328c8ca954" (UID: "41b1aa0c-723c-4f3f-9e52-26328c8ca954"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.103144 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "41b1aa0c-723c-4f3f-9e52-26328c8ca954" (UID: "41b1aa0c-723c-4f3f-9e52-26328c8ca954"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.106500 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:22:27 crc kubenswrapper[4828]: E1210 19:22:27.107192 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b1aa0c-723c-4f3f-9e52-26328c8ca954" containerName="dnsmasq-dns" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.107229 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b1aa0c-723c-4f3f-9e52-26328c8ca954" containerName="dnsmasq-dns" Dec 10 19:22:27 crc kubenswrapper[4828]: E1210 19:22:27.107247 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e369bc-b3a1-444c-a16d-e70c0b754cef" containerName="nova-manage" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.107253 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e369bc-b3a1-444c-a16d-e70c0b754cef" containerName="nova-manage" Dec 10 19:22:27 crc kubenswrapper[4828]: E1210 19:22:27.107348 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="529a0e7b-83de-4c8f-aeb6-4575868cdf74" containerName="nova-cell1-novncproxy-novncproxy" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.107357 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="529a0e7b-83de-4c8f-aeb6-4575868cdf74" containerName="nova-cell1-novncproxy-novncproxy" Dec 10 19:22:27 crc kubenswrapper[4828]: E1210 19:22:27.107369 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41b1aa0c-723c-4f3f-9e52-26328c8ca954" containerName="init" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.107374 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="41b1aa0c-723c-4f3f-9e52-26328c8ca954" containerName="init" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.107893 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5e369bc-b3a1-444c-a16d-e70c0b754cef" containerName="nova-manage" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.107928 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="41b1aa0c-723c-4f3f-9e52-26328c8ca954" containerName="dnsmasq-dns" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.107955 4828 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.107980 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.107993 4828 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.108006 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.107982 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="529a0e7b-83de-4c8f-aeb6-4575868cdf74" containerName="nova-cell1-novncproxy-novncproxy" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.109185 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.112331 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.112508 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.114215 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.115248 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "41b1aa0c-723c-4f3f-9e52-26328c8ca954" (UID: "41b1aa0c-723c-4f3f-9e52-26328c8ca954"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.122119 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.210063 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/78c25c19-441e-4b49-82b6-cad0ff109175-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"78c25c19-441e-4b49-82b6-cad0ff109175\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.210185 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/78c25c19-441e-4b49-82b6-cad0ff109175-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"78c25c19-441e-4b49-82b6-cad0ff109175\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.210301 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c25c19-441e-4b49-82b6-cad0ff109175-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"78c25c19-441e-4b49-82b6-cad0ff109175\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.210321 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c25c19-441e-4b49-82b6-cad0ff109175-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"78c25c19-441e-4b49-82b6-cad0ff109175\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.210374 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zf69\" (UniqueName: \"kubernetes.io/projected/78c25c19-441e-4b49-82b6-cad0ff109175-kube-api-access-8zf69\") pod \"nova-cell1-novncproxy-0\" (UID: \"78c25c19-441e-4b49-82b6-cad0ff109175\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.210485 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/41b1aa0c-723c-4f3f-9e52-26328c8ca954-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.247473 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-jlhgg"] Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.258108 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-jlhgg"] Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.312388 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c25c19-441e-4b49-82b6-cad0ff109175-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"78c25c19-441e-4b49-82b6-cad0ff109175\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.312432 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c25c19-441e-4b49-82b6-cad0ff109175-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"78c25c19-441e-4b49-82b6-cad0ff109175\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.312486 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zf69\" (UniqueName: \"kubernetes.io/projected/78c25c19-441e-4b49-82b6-cad0ff109175-kube-api-access-8zf69\") pod \"nova-cell1-novncproxy-0\" (UID: \"78c25c19-441e-4b49-82b6-cad0ff109175\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.312554 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/78c25c19-441e-4b49-82b6-cad0ff109175-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"78c25c19-441e-4b49-82b6-cad0ff109175\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.312632 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/78c25c19-441e-4b49-82b6-cad0ff109175-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"78c25c19-441e-4b49-82b6-cad0ff109175\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.318522 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78c25c19-441e-4b49-82b6-cad0ff109175-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"78c25c19-441e-4b49-82b6-cad0ff109175\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.318944 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/78c25c19-441e-4b49-82b6-cad0ff109175-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"78c25c19-441e-4b49-82b6-cad0ff109175\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.321048 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/78c25c19-441e-4b49-82b6-cad0ff109175-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"78c25c19-441e-4b49-82b6-cad0ff109175\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.325574 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78c25c19-441e-4b49-82b6-cad0ff109175-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"78c25c19-441e-4b49-82b6-cad0ff109175\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.335397 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zf69\" (UniqueName: \"kubernetes.io/projected/78c25c19-441e-4b49-82b6-cad0ff109175-kube-api-access-8zf69\") pod \"nova-cell1-novncproxy-0\" (UID: \"78c25c19-441e-4b49-82b6-cad0ff109175\") " pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.435412 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.566948 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.567245 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="366eb842-07b0-456f-95c2-d5f93556def8" containerName="nova-api-log" containerID="cri-o://240b44e4a825640b5f5b51f3b697b1f15462d0b1d7cbfcdff74f08f7e5863ed7" gracePeriod=30 Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.567872 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="366eb842-07b0-456f-95c2-d5f93556def8" containerName="nova-api-api" containerID="cri-o://772a8274bb34ae1d80fcd1523e509e5a334556f577dc97317ed8e41d2d84e955" gracePeriod=30 Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.594584 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.594790 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="86fca083-3b11-4d59-9a0c-8099d8257622" containerName="nova-scheduler-scheduler" containerID="cri-o://183616e4449fa53ec1ca869e092100729984d6f74eead5b35b19268e7cef0654" gracePeriod=30 Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.595310 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="366eb842-07b0-456f-95c2-d5f93556def8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.231:8774/\": EOF" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.608539 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 19:22:27 crc kubenswrapper[4828]: E1210 19:22:27.619416 4828 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="183616e4449fa53ec1ca869e092100729984d6f74eead5b35b19268e7cef0654" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 19:22:27 crc kubenswrapper[4828]: E1210 19:22:27.628593 4828 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="183616e4449fa53ec1ca869e092100729984d6f74eead5b35b19268e7cef0654" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 19:22:27 crc kubenswrapper[4828]: E1210 19:22:27.631132 4828 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="183616e4449fa53ec1ca869e092100729984d6f74eead5b35b19268e7cef0654" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 10 19:22:27 crc kubenswrapper[4828]: E1210 19:22:27.631166 4828 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="86fca083-3b11-4d59-9a0c-8099d8257622" containerName="nova-scheduler-scheduler" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.811468 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41b1aa0c-723c-4f3f-9e52-26328c8ca954" path="/var/lib/kubelet/pods/41b1aa0c-723c-4f3f-9e52-26328c8ca954/volumes" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.812648 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="529a0e7b-83de-4c8f-aeb6-4575868cdf74" path="/var/lib/kubelet/pods/529a0e7b-83de-4c8f-aeb6-4575868cdf74/volumes" Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.946541 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlq5k" event={"ID":"a615a974-7cc4-4ed1-957f-9ca4f1024f26","Type":"ContainerStarted","Data":"e76ec52e5c5c6052d67434b79ee9be8bb4a5f4d63f76034aed53ca5f4a7eb370"} Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.956763 4828 generic.go:334] "Generic (PLEG): container finished" podID="366eb842-07b0-456f-95c2-d5f93556def8" containerID="240b44e4a825640b5f5b51f3b697b1f15462d0b1d7cbfcdff74f08f7e5863ed7" exitCode=143 Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.956853 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"366eb842-07b0-456f-95c2-d5f93556def8","Type":"ContainerDied","Data":"240b44e4a825640b5f5b51f3b697b1f15462d0b1d7cbfcdff74f08f7e5863ed7"} Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.968346 4828 generic.go:334] "Generic (PLEG): container finished" podID="07160fa2-a003-4436-a117-b79d5f5627c4" containerID="f4b05950edd0b0b24d7f0423d2bc8320cfbdfcfa3c56d4df9a31fdce6993c881" exitCode=0 Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.968439 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07160fa2-a003-4436-a117-b79d5f5627c4","Type":"ContainerDied","Data":"f4b05950edd0b0b24d7f0423d2bc8320cfbdfcfa3c56d4df9a31fdce6993c881"} Dec 10 19:22:27 crc kubenswrapper[4828]: I1210 19:22:27.978561 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xlq5k" podStartSLOduration=3.944773571 podStartE2EDuration="13.978536147s" podCreationTimestamp="2025-12-10 19:22:14 +0000 UTC" firstStartedPulling="2025-12-10 19:22:16.705159751 +0000 UTC m=+1617.215770756" lastFinishedPulling="2025-12-10 19:22:26.738922327 +0000 UTC m=+1627.249533332" observedRunningTime="2025-12-10 19:22:27.965629655 +0000 UTC m=+1628.476240660" watchObservedRunningTime="2025-12-10 19:22:27.978536147 +0000 UTC m=+1628.489147152" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.051314 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.207673 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.348230 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07160fa2-a003-4436-a117-b79d5f5627c4-combined-ca-bundle\") pod \"07160fa2-a003-4436-a117-b79d5f5627c4\" (UID: \"07160fa2-a003-4436-a117-b79d5f5627c4\") " Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.349100 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07160fa2-a003-4436-a117-b79d5f5627c4-config-data\") pod \"07160fa2-a003-4436-a117-b79d5f5627c4\" (UID: \"07160fa2-a003-4436-a117-b79d5f5627c4\") " Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.349259 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07160fa2-a003-4436-a117-b79d5f5627c4-logs\") pod \"07160fa2-a003-4436-a117-b79d5f5627c4\" (UID: \"07160fa2-a003-4436-a117-b79d5f5627c4\") " Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.349334 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2r8j\" (UniqueName: \"kubernetes.io/projected/07160fa2-a003-4436-a117-b79d5f5627c4-kube-api-access-d2r8j\") pod \"07160fa2-a003-4436-a117-b79d5f5627c4\" (UID: \"07160fa2-a003-4436-a117-b79d5f5627c4\") " Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.349955 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07160fa2-a003-4436-a117-b79d5f5627c4-logs" (OuterVolumeSpecName: "logs") pod "07160fa2-a003-4436-a117-b79d5f5627c4" (UID: "07160fa2-a003-4436-a117-b79d5f5627c4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.351280 4828 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07160fa2-a003-4436-a117-b79d5f5627c4-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.353488 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07160fa2-a003-4436-a117-b79d5f5627c4-kube-api-access-d2r8j" (OuterVolumeSpecName: "kube-api-access-d2r8j") pod "07160fa2-a003-4436-a117-b79d5f5627c4" (UID: "07160fa2-a003-4436-a117-b79d5f5627c4"). InnerVolumeSpecName "kube-api-access-d2r8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.383969 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07160fa2-a003-4436-a117-b79d5f5627c4-config-data" (OuterVolumeSpecName: "config-data") pod "07160fa2-a003-4436-a117-b79d5f5627c4" (UID: "07160fa2-a003-4436-a117-b79d5f5627c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.387465 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07160fa2-a003-4436-a117-b79d5f5627c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "07160fa2-a003-4436-a117-b79d5f5627c4" (UID: "07160fa2-a003-4436-a117-b79d5f5627c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.453543 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07160fa2-a003-4436-a117-b79d5f5627c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.453585 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07160fa2-a003-4436-a117-b79d5f5627c4-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.453597 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2r8j\" (UniqueName: \"kubernetes.io/projected/07160fa2-a003-4436-a117-b79d5f5627c4-kube-api-access-d2r8j\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.485868 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-wr6ll" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.657092 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-scripts\") pod \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\" (UID: \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\") " Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.657198 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-combined-ca-bundle\") pod \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\" (UID: \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\") " Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.657343 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-config-data\") pod \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\" (UID: \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\") " Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.657406 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89t5p\" (UniqueName: \"kubernetes.io/projected/f0793df8-d37f-4645-8f1d-3cc501d1ce40-kube-api-access-89t5p\") pod \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\" (UID: \"f0793df8-d37f-4645-8f1d-3cc501d1ce40\") " Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.661247 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0793df8-d37f-4645-8f1d-3cc501d1ce40-kube-api-access-89t5p" (OuterVolumeSpecName: "kube-api-access-89t5p") pod "f0793df8-d37f-4645-8f1d-3cc501d1ce40" (UID: "f0793df8-d37f-4645-8f1d-3cc501d1ce40"). InnerVolumeSpecName "kube-api-access-89t5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.669478 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-scripts" (OuterVolumeSpecName: "scripts") pod "f0793df8-d37f-4645-8f1d-3cc501d1ce40" (UID: "f0793df8-d37f-4645-8f1d-3cc501d1ce40"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.701953 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0793df8-d37f-4645-8f1d-3cc501d1ce40" (UID: "f0793df8-d37f-4645-8f1d-3cc501d1ce40"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.705452 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-config-data" (OuterVolumeSpecName: "config-data") pod "f0793df8-d37f-4645-8f1d-3cc501d1ce40" (UID: "f0793df8-d37f-4645-8f1d-3cc501d1ce40"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.760308 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.760340 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.760349 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89t5p\" (UniqueName: \"kubernetes.io/projected/f0793df8-d37f-4645-8f1d-3cc501d1ce40-kube-api-access-89t5p\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:28 crc kubenswrapper[4828]: I1210 19:22:28.760358 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0793df8-d37f-4645-8f1d-3cc501d1ce40-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.029062 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"07160fa2-a003-4436-a117-b79d5f5627c4","Type":"ContainerDied","Data":"979c0065a7464630a78151441bd02aab96731c1cbbfb54d5bb8ac6f006323dd8"} Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.029121 4828 scope.go:117] "RemoveContainer" containerID="f4b05950edd0b0b24d7f0423d2bc8320cfbdfcfa3c56d4df9a31fdce6993c881" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.029274 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.045988 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"78c25c19-441e-4b49-82b6-cad0ff109175","Type":"ContainerStarted","Data":"87ae2540b99e3edb8c5f0295741f67ba58f03d3e54c788bcd5ac23bac2f0c8e0"} Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.046049 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 10 19:22:29 crc kubenswrapper[4828]: E1210 19:22:29.046559 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0793df8-d37f-4645-8f1d-3cc501d1ce40" containerName="nova-cell1-conductor-db-sync" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.046578 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0793df8-d37f-4645-8f1d-3cc501d1ce40" containerName="nova-cell1-conductor-db-sync" Dec 10 19:22:29 crc kubenswrapper[4828]: E1210 19:22:29.046596 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07160fa2-a003-4436-a117-b79d5f5627c4" containerName="nova-metadata-log" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.046602 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="07160fa2-a003-4436-a117-b79d5f5627c4" containerName="nova-metadata-log" Dec 10 19:22:29 crc kubenswrapper[4828]: E1210 19:22:29.046609 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07160fa2-a003-4436-a117-b79d5f5627c4" containerName="nova-metadata-metadata" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.046615 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="07160fa2-a003-4436-a117-b79d5f5627c4" containerName="nova-metadata-metadata" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.046954 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="07160fa2-a003-4436-a117-b79d5f5627c4" containerName="nova-metadata-log" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.046984 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="07160fa2-a003-4436-a117-b79d5f5627c4" containerName="nova-metadata-metadata" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.047035 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0793df8-d37f-4645-8f1d-3cc501d1ce40" containerName="nova-cell1-conductor-db-sync" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.047774 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"78c25c19-441e-4b49-82b6-cad0ff109175","Type":"ContainerStarted","Data":"19ba5e505ccf286be8f548b0c4334547a4e67554765c8917085997afb94665a5"} Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.048009 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.061410 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-wr6ll" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.061692 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-wr6ll" event={"ID":"f0793df8-d37f-4645-8f1d-3cc501d1ce40","Type":"ContainerDied","Data":"9772372298df53e4a34e99846d688a77d96081b79ea18e0fd7aea91919e460cd"} Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.061743 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9772372298df53e4a34e99846d688a77d96081b79ea18e0fd7aea91919e460cd" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.082464 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.083885 4828 scope.go:117] "RemoveContainer" containerID="e945d8ddd0409a4467a895c081dc91e12683d5639acf76aedd81ce5ddb1c306c" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.098686 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.098665012 podStartE2EDuration="3.098665012s" podCreationTimestamp="2025-12-10 19:22:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:22:29.073435664 +0000 UTC m=+1629.584046669" watchObservedRunningTime="2025-12-10 19:22:29.098665012 +0000 UTC m=+1629.609276017" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.150982 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.168311 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.168454 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79e56b96-2dd9-4c50-8440-73a2bfbba5df-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"79e56b96-2dd9-4c50-8440-73a2bfbba5df\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.168505 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79e56b96-2dd9-4c50-8440-73a2bfbba5df-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"79e56b96-2dd9-4c50-8440-73a2bfbba5df\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.168559 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptjp7\" (UniqueName: \"kubernetes.io/projected/79e56b96-2dd9-4c50-8440-73a2bfbba5df-kube-api-access-ptjp7\") pod \"nova-cell1-conductor-0\" (UID: \"79e56b96-2dd9-4c50-8440-73a2bfbba5df\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.179054 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.181096 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.183145 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.183972 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.190968 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.270738 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-config-data\") pod \"nova-metadata-0\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.270888 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k88fd\" (UniqueName: \"kubernetes.io/projected/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-kube-api-access-k88fd\") pod \"nova-metadata-0\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.270937 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.270985 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-logs\") pod \"nova-metadata-0\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.271052 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79e56b96-2dd9-4c50-8440-73a2bfbba5df-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"79e56b96-2dd9-4c50-8440-73a2bfbba5df\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.271079 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79e56b96-2dd9-4c50-8440-73a2bfbba5df-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"79e56b96-2dd9-4c50-8440-73a2bfbba5df\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.271159 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.271188 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptjp7\" (UniqueName: \"kubernetes.io/projected/79e56b96-2dd9-4c50-8440-73a2bfbba5df-kube-api-access-ptjp7\") pod \"nova-cell1-conductor-0\" (UID: \"79e56b96-2dd9-4c50-8440-73a2bfbba5df\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.275351 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79e56b96-2dd9-4c50-8440-73a2bfbba5df-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"79e56b96-2dd9-4c50-8440-73a2bfbba5df\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.277298 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/79e56b96-2dd9-4c50-8440-73a2bfbba5df-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"79e56b96-2dd9-4c50-8440-73a2bfbba5df\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.290048 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptjp7\" (UniqueName: \"kubernetes.io/projected/79e56b96-2dd9-4c50-8440-73a2bfbba5df-kube-api-access-ptjp7\") pod \"nova-cell1-conductor-0\" (UID: \"79e56b96-2dd9-4c50-8440-73a2bfbba5df\") " pod="openstack/nova-cell1-conductor-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.375675 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.375809 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-config-data\") pod \"nova-metadata-0\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.375977 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k88fd\" (UniqueName: \"kubernetes.io/projected/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-kube-api-access-k88fd\") pod \"nova-metadata-0\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.376021 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.376102 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-logs\") pod \"nova-metadata-0\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.376609 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-logs\") pod \"nova-metadata-0\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.379255 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-config-data\") pod \"nova-metadata-0\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.379435 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.379566 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.394626 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k88fd\" (UniqueName: \"kubernetes.io/projected/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-kube-api-access-k88fd\") pod \"nova-metadata-0\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.408083 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.504279 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.803340 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07160fa2-a003-4436-a117-b79d5f5627c4" path="/var/lib/kubelet/pods/07160fa2-a003-4436-a117-b79d5f5627c4/volumes" Dec 10 19:22:29 crc kubenswrapper[4828]: W1210 19:22:29.901415 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79e56b96_2dd9_4c50_8440_73a2bfbba5df.slice/crio-fa7f6709929da1e1655f978ddeb3e9c879bf73904482abc0e53a3d07d0510b99 WatchSource:0}: Error finding container fa7f6709929da1e1655f978ddeb3e9c879bf73904482abc0e53a3d07d0510b99: Status 404 returned error can't find the container with id fa7f6709929da1e1655f978ddeb3e9c879bf73904482abc0e53a3d07d0510b99 Dec 10 19:22:29 crc kubenswrapper[4828]: I1210 19:22:29.903776 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 10 19:22:30 crc kubenswrapper[4828]: I1210 19:22:30.030124 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:22:30 crc kubenswrapper[4828]: W1210 19:22:30.064865 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb6cfeff_40e0_4bb8_ac4d_55cf6043537c.slice/crio-afbdbd00890f4f43a90831d8ca85136d4eb9e1d042d728573c62c25e19334f37 WatchSource:0}: Error finding container afbdbd00890f4f43a90831d8ca85136d4eb9e1d042d728573c62c25e19334f37: Status 404 returned error can't find the container with id afbdbd00890f4f43a90831d8ca85136d4eb9e1d042d728573c62c25e19334f37 Dec 10 19:22:30 crc kubenswrapper[4828]: I1210 19:22:30.073404 4828 generic.go:334] "Generic (PLEG): container finished" podID="057a6200-117a-47be-97df-55ae94e6f763" containerID="68e5730b588bb1538a77492edf973b42c140868fae4a86ba497bcb09bbca320a" exitCode=0 Dec 10 19:22:30 crc kubenswrapper[4828]: I1210 19:22:30.073484 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-dtpxc" event={"ID":"057a6200-117a-47be-97df-55ae94e6f763","Type":"ContainerDied","Data":"68e5730b588bb1538a77492edf973b42c140868fae4a86ba497bcb09bbca320a"} Dec 10 19:22:30 crc kubenswrapper[4828]: I1210 19:22:30.074437 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"79e56b96-2dd9-4c50-8440-73a2bfbba5df","Type":"ContainerStarted","Data":"fa7f6709929da1e1655f978ddeb3e9c879bf73904482abc0e53a3d07d0510b99"} Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.112367 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"db6cfeff-40e0-4bb8-ac4d-55cf6043537c","Type":"ContainerStarted","Data":"9ac5a7c6345199c72ec2b239a21d08ad9a0bb7be487a20cdd00d11e9ba1f4e79"} Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.112642 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"db6cfeff-40e0-4bb8-ac4d-55cf6043537c","Type":"ContainerStarted","Data":"9f65f1571b18850e44d5356cad53d43aaa3286a9f054b406516f8f0bbdfff9a7"} Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.112653 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"db6cfeff-40e0-4bb8-ac4d-55cf6043537c","Type":"ContainerStarted","Data":"afbdbd00890f4f43a90831d8ca85136d4eb9e1d042d728573c62c25e19334f37"} Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.115485 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"79e56b96-2dd9-4c50-8440-73a2bfbba5df","Type":"ContainerStarted","Data":"bf9d51714fd6aa7e48e60f4c6ee93198b44866f964e2998c3b7a6a9ede3c5e21"} Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.115721 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.120883 4828 generic.go:334] "Generic (PLEG): container finished" podID="366eb842-07b0-456f-95c2-d5f93556def8" containerID="772a8274bb34ae1d80fcd1523e509e5a334556f577dc97317ed8e41d2d84e955" exitCode=0 Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.120957 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"366eb842-07b0-456f-95c2-d5f93556def8","Type":"ContainerDied","Data":"772a8274bb34ae1d80fcd1523e509e5a334556f577dc97317ed8e41d2d84e955"} Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.139359 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.139336368 podStartE2EDuration="2.139336368s" podCreationTimestamp="2025-12-10 19:22:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:22:31.132710642 +0000 UTC m=+1631.643321667" watchObservedRunningTime="2025-12-10 19:22:31.139336368 +0000 UTC m=+1631.649947373" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.157301 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.157281813 podStartE2EDuration="2.157281813s" podCreationTimestamp="2025-12-10 19:22:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:22:31.153491603 +0000 UTC m=+1631.664102618" watchObservedRunningTime="2025-12-10 19:22:31.157281813 +0000 UTC m=+1631.667892818" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.253301 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.440020 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/366eb842-07b0-456f-95c2-d5f93556def8-logs\") pod \"366eb842-07b0-456f-95c2-d5f93556def8\" (UID: \"366eb842-07b0-456f-95c2-d5f93556def8\") " Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.440220 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366eb842-07b0-456f-95c2-d5f93556def8-combined-ca-bundle\") pod \"366eb842-07b0-456f-95c2-d5f93556def8\" (UID: \"366eb842-07b0-456f-95c2-d5f93556def8\") " Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.440262 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/366eb842-07b0-456f-95c2-d5f93556def8-config-data\") pod \"366eb842-07b0-456f-95c2-d5f93556def8\" (UID: \"366eb842-07b0-456f-95c2-d5f93556def8\") " Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.440398 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjxqt\" (UniqueName: \"kubernetes.io/projected/366eb842-07b0-456f-95c2-d5f93556def8-kube-api-access-hjxqt\") pod \"366eb842-07b0-456f-95c2-d5f93556def8\" (UID: \"366eb842-07b0-456f-95c2-d5f93556def8\") " Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.440531 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/366eb842-07b0-456f-95c2-d5f93556def8-logs" (OuterVolumeSpecName: "logs") pod "366eb842-07b0-456f-95c2-d5f93556def8" (UID: "366eb842-07b0-456f-95c2-d5f93556def8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.441215 4828 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/366eb842-07b0-456f-95c2-d5f93556def8-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.445578 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/366eb842-07b0-456f-95c2-d5f93556def8-kube-api-access-hjxqt" (OuterVolumeSpecName: "kube-api-access-hjxqt") pod "366eb842-07b0-456f-95c2-d5f93556def8" (UID: "366eb842-07b0-456f-95c2-d5f93556def8"). InnerVolumeSpecName "kube-api-access-hjxqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.472855 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/366eb842-07b0-456f-95c2-d5f93556def8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "366eb842-07b0-456f-95c2-d5f93556def8" (UID: "366eb842-07b0-456f-95c2-d5f93556def8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.477000 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/366eb842-07b0-456f-95c2-d5f93556def8-config-data" (OuterVolumeSpecName: "config-data") pod "366eb842-07b0-456f-95c2-d5f93556def8" (UID: "366eb842-07b0-456f-95c2-d5f93556def8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.543721 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjxqt\" (UniqueName: \"kubernetes.io/projected/366eb842-07b0-456f-95c2-d5f93556def8-kube-api-access-hjxqt\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.543761 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/366eb842-07b0-456f-95c2-d5f93556def8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.543774 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/366eb842-07b0-456f-95c2-d5f93556def8-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.684841 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-dtpxc" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.851356 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-combined-ca-bundle\") pod \"057a6200-117a-47be-97df-55ae94e6f763\" (UID: \"057a6200-117a-47be-97df-55ae94e6f763\") " Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.851449 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-scripts\") pod \"057a6200-117a-47be-97df-55ae94e6f763\" (UID: \"057a6200-117a-47be-97df-55ae94e6f763\") " Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.851581 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzjlw\" (UniqueName: \"kubernetes.io/projected/057a6200-117a-47be-97df-55ae94e6f763-kube-api-access-vzjlw\") pod \"057a6200-117a-47be-97df-55ae94e6f763\" (UID: \"057a6200-117a-47be-97df-55ae94e6f763\") " Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.851759 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-config-data\") pod \"057a6200-117a-47be-97df-55ae94e6f763\" (UID: \"057a6200-117a-47be-97df-55ae94e6f763\") " Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.859438 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-scripts" (OuterVolumeSpecName: "scripts") pod "057a6200-117a-47be-97df-55ae94e6f763" (UID: "057a6200-117a-47be-97df-55ae94e6f763"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.859511 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/057a6200-117a-47be-97df-55ae94e6f763-kube-api-access-vzjlw" (OuterVolumeSpecName: "kube-api-access-vzjlw") pod "057a6200-117a-47be-97df-55ae94e6f763" (UID: "057a6200-117a-47be-97df-55ae94e6f763"). InnerVolumeSpecName "kube-api-access-vzjlw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.888107 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-config-data" (OuterVolumeSpecName: "config-data") pod "057a6200-117a-47be-97df-55ae94e6f763" (UID: "057a6200-117a-47be-97df-55ae94e6f763"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.889383 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "057a6200-117a-47be-97df-55ae94e6f763" (UID: "057a6200-117a-47be-97df-55ae94e6f763"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.955247 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.955292 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.955312 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/057a6200-117a-47be-97df-55ae94e6f763-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.955324 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzjlw\" (UniqueName: \"kubernetes.io/projected/057a6200-117a-47be-97df-55ae94e6f763-kube-api-access-vzjlw\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:31 crc kubenswrapper[4828]: I1210 19:22:31.977396 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.133442 4828 generic.go:334] "Generic (PLEG): container finished" podID="86fca083-3b11-4d59-9a0c-8099d8257622" containerID="183616e4449fa53ec1ca869e092100729984d6f74eead5b35b19268e7cef0654" exitCode=0 Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.133480 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"86fca083-3b11-4d59-9a0c-8099d8257622","Type":"ContainerDied","Data":"183616e4449fa53ec1ca869e092100729984d6f74eead5b35b19268e7cef0654"} Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.133596 4828 scope.go:117] "RemoveContainer" containerID="183616e4449fa53ec1ca869e092100729984d6f74eead5b35b19268e7cef0654" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.133895 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"86fca083-3b11-4d59-9a0c-8099d8257622","Type":"ContainerDied","Data":"0a4e1db2900bd5af9a0508698e0dccf59cda9f77465ec37af3c8bdb7015b2ccc"} Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.134709 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.136286 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-dtpxc" event={"ID":"057a6200-117a-47be-97df-55ae94e6f763","Type":"ContainerDied","Data":"64f94a1eca02b850a046b6892a2e8bcf7d297d26a5a11d0cb1ccee60f7b63e19"} Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.136319 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64f94a1eca02b850a046b6892a2e8bcf7d297d26a5a11d0cb1ccee60f7b63e19" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.136366 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-dtpxc" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.139281 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"366eb842-07b0-456f-95c2-d5f93556def8","Type":"ContainerDied","Data":"881eca32b1805f190d7ccc823c5561212a90ec51a43905b2ea42eb78ed0a525b"} Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.139391 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.159377 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gh24x\" (UniqueName: \"kubernetes.io/projected/86fca083-3b11-4d59-9a0c-8099d8257622-kube-api-access-gh24x\") pod \"86fca083-3b11-4d59-9a0c-8099d8257622\" (UID: \"86fca083-3b11-4d59-9a0c-8099d8257622\") " Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.159645 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86fca083-3b11-4d59-9a0c-8099d8257622-combined-ca-bundle\") pod \"86fca083-3b11-4d59-9a0c-8099d8257622\" (UID: \"86fca083-3b11-4d59-9a0c-8099d8257622\") " Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.159742 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86fca083-3b11-4d59-9a0c-8099d8257622-config-data\") pod \"86fca083-3b11-4d59-9a0c-8099d8257622\" (UID: \"86fca083-3b11-4d59-9a0c-8099d8257622\") " Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.169471 4828 scope.go:117] "RemoveContainer" containerID="183616e4449fa53ec1ca869e092100729984d6f74eead5b35b19268e7cef0654" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.170463 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86fca083-3b11-4d59-9a0c-8099d8257622-kube-api-access-gh24x" (OuterVolumeSpecName: "kube-api-access-gh24x") pod "86fca083-3b11-4d59-9a0c-8099d8257622" (UID: "86fca083-3b11-4d59-9a0c-8099d8257622"). InnerVolumeSpecName "kube-api-access-gh24x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:32 crc kubenswrapper[4828]: E1210 19:22:32.174098 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"183616e4449fa53ec1ca869e092100729984d6f74eead5b35b19268e7cef0654\": container with ID starting with 183616e4449fa53ec1ca869e092100729984d6f74eead5b35b19268e7cef0654 not found: ID does not exist" containerID="183616e4449fa53ec1ca869e092100729984d6f74eead5b35b19268e7cef0654" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.174157 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"183616e4449fa53ec1ca869e092100729984d6f74eead5b35b19268e7cef0654"} err="failed to get container status \"183616e4449fa53ec1ca869e092100729984d6f74eead5b35b19268e7cef0654\": rpc error: code = NotFound desc = could not find container \"183616e4449fa53ec1ca869e092100729984d6f74eead5b35b19268e7cef0654\": container with ID starting with 183616e4449fa53ec1ca869e092100729984d6f74eead5b35b19268e7cef0654 not found: ID does not exist" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.174193 4828 scope.go:117] "RemoveContainer" containerID="772a8274bb34ae1d80fcd1523e509e5a334556f577dc97317ed8e41d2d84e955" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.201975 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.208850 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86fca083-3b11-4d59-9a0c-8099d8257622-config-data" (OuterVolumeSpecName: "config-data") pod "86fca083-3b11-4d59-9a0c-8099d8257622" (UID: "86fca083-3b11-4d59-9a0c-8099d8257622"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.226979 4828 scope.go:117] "RemoveContainer" containerID="240b44e4a825640b5f5b51f3b697b1f15462d0b1d7cbfcdff74f08f7e5863ed7" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.235275 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.242773 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86fca083-3b11-4d59-9a0c-8099d8257622-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86fca083-3b11-4d59-9a0c-8099d8257622" (UID: "86fca083-3b11-4d59-9a0c-8099d8257622"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.245590 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 19:22:32 crc kubenswrapper[4828]: E1210 19:22:32.246126 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86fca083-3b11-4d59-9a0c-8099d8257622" containerName="nova-scheduler-scheduler" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.246140 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="86fca083-3b11-4d59-9a0c-8099d8257622" containerName="nova-scheduler-scheduler" Dec 10 19:22:32 crc kubenswrapper[4828]: E1210 19:22:32.246165 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="057a6200-117a-47be-97df-55ae94e6f763" containerName="aodh-db-sync" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.246170 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="057a6200-117a-47be-97df-55ae94e6f763" containerName="aodh-db-sync" Dec 10 19:22:32 crc kubenswrapper[4828]: E1210 19:22:32.246178 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="366eb842-07b0-456f-95c2-d5f93556def8" containerName="nova-api-log" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.246184 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="366eb842-07b0-456f-95c2-d5f93556def8" containerName="nova-api-log" Dec 10 19:22:32 crc kubenswrapper[4828]: E1210 19:22:32.246214 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="366eb842-07b0-456f-95c2-d5f93556def8" containerName="nova-api-api" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.246221 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="366eb842-07b0-456f-95c2-d5f93556def8" containerName="nova-api-api" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.246447 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="86fca083-3b11-4d59-9a0c-8099d8257622" containerName="nova-scheduler-scheduler" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.246467 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="366eb842-07b0-456f-95c2-d5f93556def8" containerName="nova-api-log" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.246487 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="057a6200-117a-47be-97df-55ae94e6f763" containerName="aodh-db-sync" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.246497 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="366eb842-07b0-456f-95c2-d5f93556def8" containerName="nova-api-api" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.247732 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.251085 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.265441 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.267387 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37407428-f199-4082-97ea-d9568eddfc1f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"37407428-f199-4082-97ea-d9568eddfc1f\") " pod="openstack/nova-api-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.267590 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37407428-f199-4082-97ea-d9568eddfc1f-config-data\") pod \"nova-api-0\" (UID: \"37407428-f199-4082-97ea-d9568eddfc1f\") " pod="openstack/nova-api-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.268177 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gntv6\" (UniqueName: \"kubernetes.io/projected/37407428-f199-4082-97ea-d9568eddfc1f-kube-api-access-gntv6\") pod \"nova-api-0\" (UID: \"37407428-f199-4082-97ea-d9568eddfc1f\") " pod="openstack/nova-api-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.268505 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37407428-f199-4082-97ea-d9568eddfc1f-logs\") pod \"nova-api-0\" (UID: \"37407428-f199-4082-97ea-d9568eddfc1f\") " pod="openstack/nova-api-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.268647 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86fca083-3b11-4d59-9a0c-8099d8257622-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.268662 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86fca083-3b11-4d59-9a0c-8099d8257622-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.268671 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gh24x\" (UniqueName: \"kubernetes.io/projected/86fca083-3b11-4d59-9a0c-8099d8257622-kube-api-access-gh24x\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.370293 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gntv6\" (UniqueName: \"kubernetes.io/projected/37407428-f199-4082-97ea-d9568eddfc1f-kube-api-access-gntv6\") pod \"nova-api-0\" (UID: \"37407428-f199-4082-97ea-d9568eddfc1f\") " pod="openstack/nova-api-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.370388 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37407428-f199-4082-97ea-d9568eddfc1f-logs\") pod \"nova-api-0\" (UID: \"37407428-f199-4082-97ea-d9568eddfc1f\") " pod="openstack/nova-api-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.370442 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37407428-f199-4082-97ea-d9568eddfc1f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"37407428-f199-4082-97ea-d9568eddfc1f\") " pod="openstack/nova-api-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.370468 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37407428-f199-4082-97ea-d9568eddfc1f-config-data\") pod \"nova-api-0\" (UID: \"37407428-f199-4082-97ea-d9568eddfc1f\") " pod="openstack/nova-api-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.370900 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37407428-f199-4082-97ea-d9568eddfc1f-logs\") pod \"nova-api-0\" (UID: \"37407428-f199-4082-97ea-d9568eddfc1f\") " pod="openstack/nova-api-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.374008 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37407428-f199-4082-97ea-d9568eddfc1f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"37407428-f199-4082-97ea-d9568eddfc1f\") " pod="openstack/nova-api-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.374242 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37407428-f199-4082-97ea-d9568eddfc1f-config-data\") pod \"nova-api-0\" (UID: \"37407428-f199-4082-97ea-d9568eddfc1f\") " pod="openstack/nova-api-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.387376 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gntv6\" (UniqueName: \"kubernetes.io/projected/37407428-f199-4082-97ea-d9568eddfc1f-kube-api-access-gntv6\") pod \"nova-api-0\" (UID: \"37407428-f199-4082-97ea-d9568eddfc1f\") " pod="openstack/nova-api-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.445201 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.544785 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.556480 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.570556 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.579917 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.582551 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.586410 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.594986 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.677251 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7c47155-fd88-4adb-962e-caf36152a6f8-config-data\") pod \"nova-scheduler-0\" (UID: \"f7c47155-fd88-4adb-962e-caf36152a6f8\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.677386 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldtrx\" (UniqueName: \"kubernetes.io/projected/f7c47155-fd88-4adb-962e-caf36152a6f8-kube-api-access-ldtrx\") pod \"nova-scheduler-0\" (UID: \"f7c47155-fd88-4adb-962e-caf36152a6f8\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.677435 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7c47155-fd88-4adb-962e-caf36152a6f8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f7c47155-fd88-4adb-962e-caf36152a6f8\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.779852 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldtrx\" (UniqueName: \"kubernetes.io/projected/f7c47155-fd88-4adb-962e-caf36152a6f8-kube-api-access-ldtrx\") pod \"nova-scheduler-0\" (UID: \"f7c47155-fd88-4adb-962e-caf36152a6f8\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.780305 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7c47155-fd88-4adb-962e-caf36152a6f8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f7c47155-fd88-4adb-962e-caf36152a6f8\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.782228 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7c47155-fd88-4adb-962e-caf36152a6f8-config-data\") pod \"nova-scheduler-0\" (UID: \"f7c47155-fd88-4adb-962e-caf36152a6f8\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.787704 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7c47155-fd88-4adb-962e-caf36152a6f8-config-data\") pod \"nova-scheduler-0\" (UID: \"f7c47155-fd88-4adb-962e-caf36152a6f8\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.793681 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7c47155-fd88-4adb-962e-caf36152a6f8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f7c47155-fd88-4adb-962e-caf36152a6f8\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.797143 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldtrx\" (UniqueName: \"kubernetes.io/projected/f7c47155-fd88-4adb-962e-caf36152a6f8-kube-api-access-ldtrx\") pod \"nova-scheduler-0\" (UID: \"f7c47155-fd88-4adb-962e-caf36152a6f8\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:32 crc kubenswrapper[4828]: I1210 19:22:32.996599 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:22:33 crc kubenswrapper[4828]: I1210 19:22:33.053293 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:22:33 crc kubenswrapper[4828]: I1210 19:22:33.158578 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37407428-f199-4082-97ea-d9568eddfc1f","Type":"ContainerStarted","Data":"1774b820f2130c959e66a06790c9b36a94985cf3faab4185bda910c44af99ead"} Dec 10 19:22:33 crc kubenswrapper[4828]: W1210 19:22:33.466914 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7c47155_fd88_4adb_962e_caf36152a6f8.slice/crio-c17bb023dfce3e592dcecbeb8b6cb1211c041959e1d7f18caf8c6a7d5d9d6d77 WatchSource:0}: Error finding container c17bb023dfce3e592dcecbeb8b6cb1211c041959e1d7f18caf8c6a7d5d9d6d77: Status 404 returned error can't find the container with id c17bb023dfce3e592dcecbeb8b6cb1211c041959e1d7f18caf8c6a7d5d9d6d77 Dec 10 19:22:33 crc kubenswrapper[4828]: I1210 19:22:33.467351 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:22:33 crc kubenswrapper[4828]: I1210 19:22:33.802507 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="366eb842-07b0-456f-95c2-d5f93556def8" path="/var/lib/kubelet/pods/366eb842-07b0-456f-95c2-d5f93556def8/volumes" Dec 10 19:22:33 crc kubenswrapper[4828]: I1210 19:22:33.803948 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86fca083-3b11-4d59-9a0c-8099d8257622" path="/var/lib/kubelet/pods/86fca083-3b11-4d59-9a0c-8099d8257622/volumes" Dec 10 19:22:33 crc kubenswrapper[4828]: I1210 19:22:33.858961 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 10 19:22:33 crc kubenswrapper[4828]: I1210 19:22:33.864365 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 10 19:22:33 crc kubenswrapper[4828]: I1210 19:22:33.867422 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 10 19:22:33 crc kubenswrapper[4828]: I1210 19:22:33.869052 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 10 19:22:33 crc kubenswrapper[4828]: I1210 19:22:33.869371 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-qg8rx" Dec 10 19:22:33 crc kubenswrapper[4828]: I1210 19:22:33.885641 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 10 19:22:33 crc kubenswrapper[4828]: I1210 19:22:33.908380 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfpw4\" (UniqueName: \"kubernetes.io/projected/60ef7c6c-dfbd-4fff-b451-80b21458f86b-kube-api-access-jfpw4\") pod \"aodh-0\" (UID: \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\") " pod="openstack/aodh-0" Dec 10 19:22:33 crc kubenswrapper[4828]: I1210 19:22:33.908936 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-config-data\") pod \"aodh-0\" (UID: \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\") " pod="openstack/aodh-0" Dec 10 19:22:33 crc kubenswrapper[4828]: I1210 19:22:33.909083 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-scripts\") pod \"aodh-0\" (UID: \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\") " pod="openstack/aodh-0" Dec 10 19:22:33 crc kubenswrapper[4828]: I1210 19:22:33.909262 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\") " pod="openstack/aodh-0" Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.016067 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfpw4\" (UniqueName: \"kubernetes.io/projected/60ef7c6c-dfbd-4fff-b451-80b21458f86b-kube-api-access-jfpw4\") pod \"aodh-0\" (UID: \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\") " pod="openstack/aodh-0" Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.016233 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-config-data\") pod \"aodh-0\" (UID: \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\") " pod="openstack/aodh-0" Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.016400 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-scripts\") pod \"aodh-0\" (UID: \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\") " pod="openstack/aodh-0" Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.018681 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\") " pod="openstack/aodh-0" Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.023308 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-config-data\") pod \"aodh-0\" (UID: \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\") " pod="openstack/aodh-0" Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.032289 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-scripts\") pod \"aodh-0\" (UID: \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\") " pod="openstack/aodh-0" Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.033808 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\") " pod="openstack/aodh-0" Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.038658 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfpw4\" (UniqueName: \"kubernetes.io/projected/60ef7c6c-dfbd-4fff-b451-80b21458f86b-kube-api-access-jfpw4\") pod \"aodh-0\" (UID: \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\") " pod="openstack/aodh-0" Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.186509 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37407428-f199-4082-97ea-d9568eddfc1f","Type":"ContainerStarted","Data":"854f52346883105f81eaf3a28670fdc042e5fa5f097525eeeb540d11f8d0c7ad"} Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.186567 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37407428-f199-4082-97ea-d9568eddfc1f","Type":"ContainerStarted","Data":"0f032ae026aa4205af0fa013963be337918313790e0b941250faba503a41d333"} Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.190396 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7c47155-fd88-4adb-962e-caf36152a6f8","Type":"ContainerStarted","Data":"4dbea57c3646137d7bc461d34f0d6c543f93a6f37639f833a0893e717b98d4d6"} Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.190423 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7c47155-fd88-4adb-962e-caf36152a6f8","Type":"ContainerStarted","Data":"c17bb023dfce3e592dcecbeb8b6cb1211c041959e1d7f18caf8c6a7d5d9d6d77"} Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.192305 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.206595 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.206576141 podStartE2EDuration="2.206576141s" podCreationTimestamp="2025-12-10 19:22:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:22:34.202481412 +0000 UTC m=+1634.713092417" watchObservedRunningTime="2025-12-10 19:22:34.206576141 +0000 UTC m=+1634.717187146" Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.217977 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.217952792 podStartE2EDuration="2.217952792s" podCreationTimestamp="2025-12-10 19:22:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:22:34.216506744 +0000 UTC m=+1634.727117759" watchObservedRunningTime="2025-12-10 19:22:34.217952792 +0000 UTC m=+1634.728563797" Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.504959 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.505264 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.586818 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xlq5k" Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.586863 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xlq5k" Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.646042 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xlq5k" Dec 10 19:22:34 crc kubenswrapper[4828]: I1210 19:22:34.684205 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 10 19:22:35 crc kubenswrapper[4828]: I1210 19:22:35.203746 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"60ef7c6c-dfbd-4fff-b451-80b21458f86b","Type":"ContainerStarted","Data":"da1b36078c5882c12ba56845d3e3c12b90ca993e6e46b564a88c2ab153a7a1c7"} Dec 10 19:22:35 crc kubenswrapper[4828]: I1210 19:22:35.266155 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xlq5k" Dec 10 19:22:35 crc kubenswrapper[4828]: I1210 19:22:35.333490 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlq5k"] Dec 10 19:22:35 crc kubenswrapper[4828]: I1210 19:22:35.789100 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:22:35 crc kubenswrapper[4828]: E1210 19:22:35.789900 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:22:36 crc kubenswrapper[4828]: I1210 19:22:36.128812 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:22:36 crc kubenswrapper[4828]: I1210 19:22:36.129176 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="ceilometer-notification-agent" containerID="cri-o://8fc37f29001dd68cd353a6955c98a84deb39e1934c5524f98a4993c0aee981ce" gracePeriod=30 Dec 10 19:22:36 crc kubenswrapper[4828]: I1210 19:22:36.129206 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="sg-core" containerID="cri-o://73bb1fae9724fc9bd5012c30e1f40798f43c89cf4eb2c0e820dbd1518eef051c" gracePeriod=30 Dec 10 19:22:36 crc kubenswrapper[4828]: I1210 19:22:36.129237 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="proxy-httpd" containerID="cri-o://5a873c23c32a5dc22e600ce16b6e6ff38985d36e40eb7913b342b61e9887f0ee" gracePeriod=30 Dec 10 19:22:36 crc kubenswrapper[4828]: I1210 19:22:36.129477 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="ceilometer-central-agent" containerID="cri-o://5c6fc4c0290724312e38ce7793e291b58430c0024bfd1582155d38594ec86ecd" gracePeriod=30 Dec 10 19:22:36 crc kubenswrapper[4828]: I1210 19:22:36.215449 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"60ef7c6c-dfbd-4fff-b451-80b21458f86b","Type":"ContainerStarted","Data":"bfc226dee05895e01f3129c274e91dca9ad880d14769124bb765b6acdfe395a4"} Dec 10 19:22:36 crc kubenswrapper[4828]: I1210 19:22:36.231061 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.238:3000/\": read tcp 10.217.0.2:38098->10.217.0.238:3000: read: connection reset by peer" Dec 10 19:22:36 crc kubenswrapper[4828]: I1210 19:22:36.982519 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.228955 4828 generic.go:334] "Generic (PLEG): container finished" podID="a8f736d0-0707-4e40-af46-4e6882b62586" containerID="5a873c23c32a5dc22e600ce16b6e6ff38985d36e40eb7913b342b61e9887f0ee" exitCode=0 Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.229444 4828 generic.go:334] "Generic (PLEG): container finished" podID="a8f736d0-0707-4e40-af46-4e6882b62586" containerID="73bb1fae9724fc9bd5012c30e1f40798f43c89cf4eb2c0e820dbd1518eef051c" exitCode=2 Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.229058 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8f736d0-0707-4e40-af46-4e6882b62586","Type":"ContainerDied","Data":"5a873c23c32a5dc22e600ce16b6e6ff38985d36e40eb7913b342b61e9887f0ee"} Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.229501 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8f736d0-0707-4e40-af46-4e6882b62586","Type":"ContainerDied","Data":"73bb1fae9724fc9bd5012c30e1f40798f43c89cf4eb2c0e820dbd1518eef051c"} Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.229518 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8f736d0-0707-4e40-af46-4e6882b62586","Type":"ContainerDied","Data":"5c6fc4c0290724312e38ce7793e291b58430c0024bfd1582155d38594ec86ecd"} Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.229457 4828 generic.go:334] "Generic (PLEG): container finished" podID="a8f736d0-0707-4e40-af46-4e6882b62586" containerID="5c6fc4c0290724312e38ce7793e291b58430c0024bfd1582155d38594ec86ecd" exitCode=0 Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.229731 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xlq5k" podUID="a615a974-7cc4-4ed1-957f-9ca4f1024f26" containerName="registry-server" containerID="cri-o://e76ec52e5c5c6052d67434b79ee9be8bb4a5f4d63f76034aed53ca5f4a7eb370" gracePeriod=2 Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.436030 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.462565 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.784860 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xlq5k" Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.900043 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a615a974-7cc4-4ed1-957f-9ca4f1024f26-utilities\") pod \"a615a974-7cc4-4ed1-957f-9ca4f1024f26\" (UID: \"a615a974-7cc4-4ed1-957f-9ca4f1024f26\") " Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.900193 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a615a974-7cc4-4ed1-957f-9ca4f1024f26-catalog-content\") pod \"a615a974-7cc4-4ed1-957f-9ca4f1024f26\" (UID: \"a615a974-7cc4-4ed1-957f-9ca4f1024f26\") " Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.900370 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpgv7\" (UniqueName: \"kubernetes.io/projected/a615a974-7cc4-4ed1-957f-9ca4f1024f26-kube-api-access-qpgv7\") pod \"a615a974-7cc4-4ed1-957f-9ca4f1024f26\" (UID: \"a615a974-7cc4-4ed1-957f-9ca4f1024f26\") " Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.900838 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a615a974-7cc4-4ed1-957f-9ca4f1024f26-utilities" (OuterVolumeSpecName: "utilities") pod "a615a974-7cc4-4ed1-957f-9ca4f1024f26" (UID: "a615a974-7cc4-4ed1-957f-9ca4f1024f26"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.901113 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a615a974-7cc4-4ed1-957f-9ca4f1024f26-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.905828 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a615a974-7cc4-4ed1-957f-9ca4f1024f26-kube-api-access-qpgv7" (OuterVolumeSpecName: "kube-api-access-qpgv7") pod "a615a974-7cc4-4ed1-957f-9ca4f1024f26" (UID: "a615a974-7cc4-4ed1-957f-9ca4f1024f26"). InnerVolumeSpecName "kube-api-access-qpgv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.926360 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a615a974-7cc4-4ed1-957f-9ca4f1024f26-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a615a974-7cc4-4ed1-957f-9ca4f1024f26" (UID: "a615a974-7cc4-4ed1-957f-9ca4f1024f26"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:22:37 crc kubenswrapper[4828]: I1210 19:22:37.997709 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.005621 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpgv7\" (UniqueName: \"kubernetes.io/projected/a615a974-7cc4-4ed1-957f-9ca4f1024f26-kube-api-access-qpgv7\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.005669 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a615a974-7cc4-4ed1-957f-9ca4f1024f26-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.248104 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"60ef7c6c-dfbd-4fff-b451-80b21458f86b","Type":"ContainerStarted","Data":"885f248b67052747290c8d222d6f14491a48c593f5c7bc9eb5f84872ea81de9a"} Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.254908 4828 generic.go:334] "Generic (PLEG): container finished" podID="a615a974-7cc4-4ed1-957f-9ca4f1024f26" containerID="e76ec52e5c5c6052d67434b79ee9be8bb4a5f4d63f76034aed53ca5f4a7eb370" exitCode=0 Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.255005 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xlq5k" Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.254998 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlq5k" event={"ID":"a615a974-7cc4-4ed1-957f-9ca4f1024f26","Type":"ContainerDied","Data":"e76ec52e5c5c6052d67434b79ee9be8bb4a5f4d63f76034aed53ca5f4a7eb370"} Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.255184 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xlq5k" event={"ID":"a615a974-7cc4-4ed1-957f-9ca4f1024f26","Type":"ContainerDied","Data":"ad09d920e335be211057ca7e61a231aefabad843081449f3556c190cbdb21e41"} Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.255217 4828 scope.go:117] "RemoveContainer" containerID="e76ec52e5c5c6052d67434b79ee9be8bb4a5f4d63f76034aed53ca5f4a7eb370" Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.271507 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.295262 4828 scope.go:117] "RemoveContainer" containerID="018b75b946941c5269e855f2e35fba21e4f6dc17513a34aa6be23cd06164f41c" Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.323925 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlq5k"] Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.334734 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xlq5k"] Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.356634 4828 scope.go:117] "RemoveContainer" containerID="406e8e47de3ab043374488bec2694a7b90ea38d4663ad1b8f2b2ee9539cbac43" Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.414090 4828 scope.go:117] "RemoveContainer" containerID="e76ec52e5c5c6052d67434b79ee9be8bb4a5f4d63f76034aed53ca5f4a7eb370" Dec 10 19:22:38 crc kubenswrapper[4828]: E1210 19:22:38.415305 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e76ec52e5c5c6052d67434b79ee9be8bb4a5f4d63f76034aed53ca5f4a7eb370\": container with ID starting with e76ec52e5c5c6052d67434b79ee9be8bb4a5f4d63f76034aed53ca5f4a7eb370 not found: ID does not exist" containerID="e76ec52e5c5c6052d67434b79ee9be8bb4a5f4d63f76034aed53ca5f4a7eb370" Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.415347 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e76ec52e5c5c6052d67434b79ee9be8bb4a5f4d63f76034aed53ca5f4a7eb370"} err="failed to get container status \"e76ec52e5c5c6052d67434b79ee9be8bb4a5f4d63f76034aed53ca5f4a7eb370\": rpc error: code = NotFound desc = could not find container \"e76ec52e5c5c6052d67434b79ee9be8bb4a5f4d63f76034aed53ca5f4a7eb370\": container with ID starting with e76ec52e5c5c6052d67434b79ee9be8bb4a5f4d63f76034aed53ca5f4a7eb370 not found: ID does not exist" Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.415373 4828 scope.go:117] "RemoveContainer" containerID="018b75b946941c5269e855f2e35fba21e4f6dc17513a34aa6be23cd06164f41c" Dec 10 19:22:38 crc kubenswrapper[4828]: E1210 19:22:38.415705 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"018b75b946941c5269e855f2e35fba21e4f6dc17513a34aa6be23cd06164f41c\": container with ID starting with 018b75b946941c5269e855f2e35fba21e4f6dc17513a34aa6be23cd06164f41c not found: ID does not exist" containerID="018b75b946941c5269e855f2e35fba21e4f6dc17513a34aa6be23cd06164f41c" Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.415746 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"018b75b946941c5269e855f2e35fba21e4f6dc17513a34aa6be23cd06164f41c"} err="failed to get container status \"018b75b946941c5269e855f2e35fba21e4f6dc17513a34aa6be23cd06164f41c\": rpc error: code = NotFound desc = could not find container \"018b75b946941c5269e855f2e35fba21e4f6dc17513a34aa6be23cd06164f41c\": container with ID starting with 018b75b946941c5269e855f2e35fba21e4f6dc17513a34aa6be23cd06164f41c not found: ID does not exist" Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.415787 4828 scope.go:117] "RemoveContainer" containerID="406e8e47de3ab043374488bec2694a7b90ea38d4663ad1b8f2b2ee9539cbac43" Dec 10 19:22:38 crc kubenswrapper[4828]: E1210 19:22:38.416145 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"406e8e47de3ab043374488bec2694a7b90ea38d4663ad1b8f2b2ee9539cbac43\": container with ID starting with 406e8e47de3ab043374488bec2694a7b90ea38d4663ad1b8f2b2ee9539cbac43 not found: ID does not exist" containerID="406e8e47de3ab043374488bec2694a7b90ea38d4663ad1b8f2b2ee9539cbac43" Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.416171 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"406e8e47de3ab043374488bec2694a7b90ea38d4663ad1b8f2b2ee9539cbac43"} err="failed to get container status \"406e8e47de3ab043374488bec2694a7b90ea38d4663ad1b8f2b2ee9539cbac43\": rpc error: code = NotFound desc = could not find container \"406e8e47de3ab043374488bec2694a7b90ea38d4663ad1b8f2b2ee9539cbac43\": container with ID starting with 406e8e47de3ab043374488bec2694a7b90ea38d4663ad1b8f2b2ee9539cbac43 not found: ID does not exist" Dec 10 19:22:38 crc kubenswrapper[4828]: I1210 19:22:38.493463 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.238:3000/\": dial tcp 10.217.0.238:3000: connect: connection refused" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.275195 4828 generic.go:334] "Generic (PLEG): container finished" podID="a8f736d0-0707-4e40-af46-4e6882b62586" containerID="8fc37f29001dd68cd353a6955c98a84deb39e1934c5524f98a4993c0aee981ce" exitCode=0 Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.275245 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8f736d0-0707-4e40-af46-4e6882b62586","Type":"ContainerDied","Data":"8fc37f29001dd68cd353a6955c98a84deb39e1934c5524f98a4993c0aee981ce"} Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.439970 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.470626 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.506109 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.506158 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.556519 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-combined-ca-bundle\") pod \"a8f736d0-0707-4e40-af46-4e6882b62586\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.556586 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-config-data\") pod \"a8f736d0-0707-4e40-af46-4e6882b62586\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.556664 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmfmd\" (UniqueName: \"kubernetes.io/projected/a8f736d0-0707-4e40-af46-4e6882b62586-kube-api-access-jmfmd\") pod \"a8f736d0-0707-4e40-af46-4e6882b62586\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.556734 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-scripts\") pod \"a8f736d0-0707-4e40-af46-4e6882b62586\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.556764 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8f736d0-0707-4e40-af46-4e6882b62586-run-httpd\") pod \"a8f736d0-0707-4e40-af46-4e6882b62586\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.556815 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8f736d0-0707-4e40-af46-4e6882b62586-log-httpd\") pod \"a8f736d0-0707-4e40-af46-4e6882b62586\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.556898 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-sg-core-conf-yaml\") pod \"a8f736d0-0707-4e40-af46-4e6882b62586\" (UID: \"a8f736d0-0707-4e40-af46-4e6882b62586\") " Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.561142 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8f736d0-0707-4e40-af46-4e6882b62586-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a8f736d0-0707-4e40-af46-4e6882b62586" (UID: "a8f736d0-0707-4e40-af46-4e6882b62586"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.561355 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a8f736d0-0707-4e40-af46-4e6882b62586-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a8f736d0-0707-4e40-af46-4e6882b62586" (UID: "a8f736d0-0707-4e40-af46-4e6882b62586"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.568276 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-scripts" (OuterVolumeSpecName: "scripts") pod "a8f736d0-0707-4e40-af46-4e6882b62586" (UID: "a8f736d0-0707-4e40-af46-4e6882b62586"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.571536 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8f736d0-0707-4e40-af46-4e6882b62586-kube-api-access-jmfmd" (OuterVolumeSpecName: "kube-api-access-jmfmd") pod "a8f736d0-0707-4e40-af46-4e6882b62586" (UID: "a8f736d0-0707-4e40-af46-4e6882b62586"). InnerVolumeSpecName "kube-api-access-jmfmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.636719 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a8f736d0-0707-4e40-af46-4e6882b62586" (UID: "a8f736d0-0707-4e40-af46-4e6882b62586"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.659425 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmfmd\" (UniqueName: \"kubernetes.io/projected/a8f736d0-0707-4e40-af46-4e6882b62586-kube-api-access-jmfmd\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.659464 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.659473 4828 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8f736d0-0707-4e40-af46-4e6882b62586-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.659482 4828 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a8f736d0-0707-4e40-af46-4e6882b62586-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.659490 4828 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.732534 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8f736d0-0707-4e40-af46-4e6882b62586" (UID: "a8f736d0-0707-4e40-af46-4e6882b62586"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.761101 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.790100 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-config-data" (OuterVolumeSpecName: "config-data") pod "a8f736d0-0707-4e40-af46-4e6882b62586" (UID: "a8f736d0-0707-4e40-af46-4e6882b62586"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.800605 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a615a974-7cc4-4ed1-957f-9ca4f1024f26" path="/var/lib/kubelet/pods/a615a974-7cc4-4ed1-957f-9ca4f1024f26/volumes" Dec 10 19:22:39 crc kubenswrapper[4828]: I1210 19:22:39.862650 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8f736d0-0707-4e40-af46-4e6882b62586-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.018251 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-v9g52"] Dec 10 19:22:40 crc kubenswrapper[4828]: E1210 19:22:40.018841 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="ceilometer-notification-agent" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.018861 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="ceilometer-notification-agent" Dec 10 19:22:40 crc kubenswrapper[4828]: E1210 19:22:40.018896 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a615a974-7cc4-4ed1-957f-9ca4f1024f26" containerName="extract-utilities" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.018905 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="a615a974-7cc4-4ed1-957f-9ca4f1024f26" containerName="extract-utilities" Dec 10 19:22:40 crc kubenswrapper[4828]: E1210 19:22:40.018919 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="proxy-httpd" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.018926 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="proxy-httpd" Dec 10 19:22:40 crc kubenswrapper[4828]: E1210 19:22:40.018940 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a615a974-7cc4-4ed1-957f-9ca4f1024f26" containerName="registry-server" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.018948 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="a615a974-7cc4-4ed1-957f-9ca4f1024f26" containerName="registry-server" Dec 10 19:22:40 crc kubenswrapper[4828]: E1210 19:22:40.018981 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a615a974-7cc4-4ed1-957f-9ca4f1024f26" containerName="extract-content" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.018989 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="a615a974-7cc4-4ed1-957f-9ca4f1024f26" containerName="extract-content" Dec 10 19:22:40 crc kubenswrapper[4828]: E1210 19:22:40.019003 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="ceilometer-central-agent" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.019010 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="ceilometer-central-agent" Dec 10 19:22:40 crc kubenswrapper[4828]: E1210 19:22:40.019025 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="sg-core" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.019032 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="sg-core" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.019304 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="ceilometer-central-agent" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.019335 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="a615a974-7cc4-4ed1-957f-9ca4f1024f26" containerName="registry-server" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.019347 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="sg-core" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.019369 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="proxy-httpd" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.019385 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" containerName="ceilometer-notification-agent" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.020433 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-v9g52" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.023812 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.024019 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.028497 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-v9g52"] Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.067867 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-config-data\") pod \"nova-cell1-cell-mapping-v9g52\" (UID: \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\") " pod="openstack/nova-cell1-cell-mapping-v9g52" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.067970 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-scripts\") pod \"nova-cell1-cell-mapping-v9g52\" (UID: \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\") " pod="openstack/nova-cell1-cell-mapping-v9g52" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.068023 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ztlt\" (UniqueName: \"kubernetes.io/projected/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-kube-api-access-5ztlt\") pod \"nova-cell1-cell-mapping-v9g52\" (UID: \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\") " pod="openstack/nova-cell1-cell-mapping-v9g52" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.068068 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-v9g52\" (UID: \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\") " pod="openstack/nova-cell1-cell-mapping-v9g52" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.169577 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-v9g52\" (UID: \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\") " pod="openstack/nova-cell1-cell-mapping-v9g52" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.170273 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-config-data\") pod \"nova-cell1-cell-mapping-v9g52\" (UID: \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\") " pod="openstack/nova-cell1-cell-mapping-v9g52" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.170450 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-scripts\") pod \"nova-cell1-cell-mapping-v9g52\" (UID: \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\") " pod="openstack/nova-cell1-cell-mapping-v9g52" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.170579 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ztlt\" (UniqueName: \"kubernetes.io/projected/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-kube-api-access-5ztlt\") pod \"nova-cell1-cell-mapping-v9g52\" (UID: \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\") " pod="openstack/nova-cell1-cell-mapping-v9g52" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.181768 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-config-data\") pod \"nova-cell1-cell-mapping-v9g52\" (UID: \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\") " pod="openstack/nova-cell1-cell-mapping-v9g52" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.185029 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-scripts\") pod \"nova-cell1-cell-mapping-v9g52\" (UID: \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\") " pod="openstack/nova-cell1-cell-mapping-v9g52" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.190273 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-v9g52\" (UID: \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\") " pod="openstack/nova-cell1-cell-mapping-v9g52" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.191381 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ztlt\" (UniqueName: \"kubernetes.io/projected/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-kube-api-access-5ztlt\") pod \"nova-cell1-cell-mapping-v9g52\" (UID: \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\") " pod="openstack/nova-cell1-cell-mapping-v9g52" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.305527 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a8f736d0-0707-4e40-af46-4e6882b62586","Type":"ContainerDied","Data":"2412de5b4686b2fa229ec4beccce1824f81aeca50d9cc7bc1401b2cff980b66a"} Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.305759 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.305929 4828 scope.go:117] "RemoveContainer" containerID="5a873c23c32a5dc22e600ce16b6e6ff38985d36e40eb7913b342b61e9887f0ee" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.311763 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"60ef7c6c-dfbd-4fff-b451-80b21458f86b","Type":"ContainerStarted","Data":"6cb044f2dfd191d9d7014f07cf4aa79d64614a633efcab0bf69681601208db5e"} Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.341605 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-v9g52" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.361127 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.377736 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.390408 4828 scope.go:117] "RemoveContainer" containerID="73bb1fae9724fc9bd5012c30e1f40798f43c89cf4eb2c0e820dbd1518eef051c" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.397881 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.402755 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.405517 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.405740 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.413732 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.478416 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g96p\" (UniqueName: \"kubernetes.io/projected/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-kube-api-access-8g96p\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.478472 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.478618 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-config-data\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.478770 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-run-httpd\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.478847 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.479107 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-scripts\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.479154 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-log-httpd\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.519984 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="db6cfeff-40e0-4bb8-ac4d-55cf6043537c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.245:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.520308 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="db6cfeff-40e0-4bb8-ac4d-55cf6043537c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.245:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.582227 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-scripts\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.582733 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-log-httpd\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.582854 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g96p\" (UniqueName: \"kubernetes.io/projected/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-kube-api-access-8g96p\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.582914 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.582991 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-config-data\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.583076 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-run-httpd\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.583110 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.590919 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.591301 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-run-httpd\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.591672 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-log-httpd\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.592454 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-config-data\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.592514 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-scripts\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.595366 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.610624 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g96p\" (UniqueName: \"kubernetes.io/projected/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-kube-api-access-8g96p\") pod \"ceilometer-0\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.799545 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:22:40 crc kubenswrapper[4828]: I1210 19:22:40.900307 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-v9g52"] Dec 10 19:22:41 crc kubenswrapper[4828]: I1210 19:22:41.485060 4828 scope.go:117] "RemoveContainer" containerID="8fc37f29001dd68cd353a6955c98a84deb39e1934c5524f98a4993c0aee981ce" Dec 10 19:22:41 crc kubenswrapper[4828]: I1210 19:22:41.789536 4828 scope.go:117] "RemoveContainer" containerID="5c6fc4c0290724312e38ce7793e291b58430c0024bfd1582155d38594ec86ecd" Dec 10 19:22:41 crc kubenswrapper[4828]: I1210 19:22:41.829785 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8f736d0-0707-4e40-af46-4e6882b62586" path="/var/lib/kubelet/pods/a8f736d0-0707-4e40-af46-4e6882b62586/volumes" Dec 10 19:22:42 crc kubenswrapper[4828]: I1210 19:22:42.094542 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:22:42 crc kubenswrapper[4828]: I1210 19:22:42.364625 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9cbf2a7-b956-4bf8-90f9-1987aafe08df","Type":"ContainerStarted","Data":"7a0dcf1298239fa559381d73eb438f4eb56c57eadd348ee76ce217b269885847"} Dec 10 19:22:42 crc kubenswrapper[4828]: I1210 19:22:42.366742 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"60ef7c6c-dfbd-4fff-b451-80b21458f86b","Type":"ContainerStarted","Data":"bc200df36df7c83b3e98bfd8302854d7b057f40b4b3c3aff31ae44ecb938792a"} Dec 10 19:22:42 crc kubenswrapper[4828]: I1210 19:22:42.366995 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerName="aodh-api" containerID="cri-o://bfc226dee05895e01f3129c274e91dca9ad880d14769124bb765b6acdfe395a4" gracePeriod=30 Dec 10 19:22:42 crc kubenswrapper[4828]: I1210 19:22:42.367084 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerName="aodh-notifier" containerID="cri-o://6cb044f2dfd191d9d7014f07cf4aa79d64614a633efcab0bf69681601208db5e" gracePeriod=30 Dec 10 19:22:42 crc kubenswrapper[4828]: I1210 19:22:42.367134 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerName="aodh-evaluator" containerID="cri-o://885f248b67052747290c8d222d6f14491a48c593f5c7bc9eb5f84872ea81de9a" gracePeriod=30 Dec 10 19:22:42 crc kubenswrapper[4828]: I1210 19:22:42.367083 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerName="aodh-listener" containerID="cri-o://bc200df36df7c83b3e98bfd8302854d7b057f40b4b3c3aff31ae44ecb938792a" gracePeriod=30 Dec 10 19:22:42 crc kubenswrapper[4828]: I1210 19:22:42.369212 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-v9g52" event={"ID":"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f","Type":"ContainerStarted","Data":"069149e5295320fff42b4e6b3c3d6f2c88abc48d4b319290aca58392c560dc3b"} Dec 10 19:22:42 crc kubenswrapper[4828]: I1210 19:22:42.369250 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-v9g52" event={"ID":"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f","Type":"ContainerStarted","Data":"74ad506bf9615e41e868725c2a8d2f29cb9c2e4510033adffa492c810e6c1f4e"} Dec 10 19:22:42 crc kubenswrapper[4828]: I1210 19:22:42.409105 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.51141612 podStartE2EDuration="9.409087598s" podCreationTimestamp="2025-12-10 19:22:33 +0000 UTC" firstStartedPulling="2025-12-10 19:22:34.692698026 +0000 UTC m=+1635.203309031" lastFinishedPulling="2025-12-10 19:22:41.590369504 +0000 UTC m=+1642.100980509" observedRunningTime="2025-12-10 19:22:42.403981062 +0000 UTC m=+1642.914592067" watchObservedRunningTime="2025-12-10 19:22:42.409087598 +0000 UTC m=+1642.919698603" Dec 10 19:22:42 crc kubenswrapper[4828]: I1210 19:22:42.435739 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-v9g52" podStartSLOduration=3.435717193 podStartE2EDuration="3.435717193s" podCreationTimestamp="2025-12-10 19:22:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:22:42.425617075 +0000 UTC m=+1642.936228090" watchObservedRunningTime="2025-12-10 19:22:42.435717193 +0000 UTC m=+1642.946328198" Dec 10 19:22:42 crc kubenswrapper[4828]: I1210 19:22:42.571872 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 19:22:42 crc kubenswrapper[4828]: I1210 19:22:42.571937 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 19:22:42 crc kubenswrapper[4828]: I1210 19:22:42.997743 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 10 19:22:43 crc kubenswrapper[4828]: I1210 19:22:43.032940 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 10 19:22:43 crc kubenswrapper[4828]: I1210 19:22:43.379863 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9cbf2a7-b956-4bf8-90f9-1987aafe08df","Type":"ContainerStarted","Data":"46bb815020c90cd623232c5f1e70bc56377d205a168df279124bb5cacf240003"} Dec 10 19:22:43 crc kubenswrapper[4828]: I1210 19:22:43.382643 4828 generic.go:334] "Generic (PLEG): container finished" podID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerID="6cb044f2dfd191d9d7014f07cf4aa79d64614a633efcab0bf69681601208db5e" exitCode=0 Dec 10 19:22:43 crc kubenswrapper[4828]: I1210 19:22:43.382674 4828 generic.go:334] "Generic (PLEG): container finished" podID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerID="885f248b67052747290c8d222d6f14491a48c593f5c7bc9eb5f84872ea81de9a" exitCode=0 Dec 10 19:22:43 crc kubenswrapper[4828]: I1210 19:22:43.382681 4828 generic.go:334] "Generic (PLEG): container finished" podID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerID="bfc226dee05895e01f3129c274e91dca9ad880d14769124bb765b6acdfe395a4" exitCode=0 Dec 10 19:22:43 crc kubenswrapper[4828]: I1210 19:22:43.382716 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"60ef7c6c-dfbd-4fff-b451-80b21458f86b","Type":"ContainerDied","Data":"6cb044f2dfd191d9d7014f07cf4aa79d64614a633efcab0bf69681601208db5e"} Dec 10 19:22:43 crc kubenswrapper[4828]: I1210 19:22:43.382755 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"60ef7c6c-dfbd-4fff-b451-80b21458f86b","Type":"ContainerDied","Data":"885f248b67052747290c8d222d6f14491a48c593f5c7bc9eb5f84872ea81de9a"} Dec 10 19:22:43 crc kubenswrapper[4828]: I1210 19:22:43.382766 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"60ef7c6c-dfbd-4fff-b451-80b21458f86b","Type":"ContainerDied","Data":"bfc226dee05895e01f3129c274e91dca9ad880d14769124bb765b6acdfe395a4"} Dec 10 19:22:43 crc kubenswrapper[4828]: I1210 19:22:43.423669 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 10 19:22:43 crc kubenswrapper[4828]: I1210 19:22:43.613270 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="37407428-f199-4082-97ea-d9568eddfc1f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.246:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 19:22:43 crc kubenswrapper[4828]: I1210 19:22:43.654061 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="37407428-f199-4082-97ea-d9568eddfc1f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.246:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 19:22:44 crc kubenswrapper[4828]: I1210 19:22:44.399383 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9cbf2a7-b956-4bf8-90f9-1987aafe08df","Type":"ContainerStarted","Data":"332a9ba66bc89a6c7a57b8052cb0fe701f866bb29da3b489d148e7c37821e1bb"} Dec 10 19:22:46 crc kubenswrapper[4828]: I1210 19:22:46.428099 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9cbf2a7-b956-4bf8-90f9-1987aafe08df","Type":"ContainerStarted","Data":"757b304b97d21f6ae4063f6a0334b634076c035c88a67045f183b3997fa16c51"} Dec 10 19:22:47 crc kubenswrapper[4828]: I1210 19:22:47.451992 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9cbf2a7-b956-4bf8-90f9-1987aafe08df","Type":"ContainerStarted","Data":"ed4bde243a6be79ae2b68d87435d741c6a9953ed37a775ae40496dbf028d88ac"} Dec 10 19:22:47 crc kubenswrapper[4828]: I1210 19:22:47.452686 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 19:22:47 crc kubenswrapper[4828]: I1210 19:22:47.457822 4828 generic.go:334] "Generic (PLEG): container finished" podID="ac39304a-7ef4-42a7-a9b3-9ea15a237a4f" containerID="069149e5295320fff42b4e6b3c3d6f2c88abc48d4b319290aca58392c560dc3b" exitCode=0 Dec 10 19:22:47 crc kubenswrapper[4828]: I1210 19:22:47.457868 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-v9g52" event={"ID":"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f","Type":"ContainerDied","Data":"069149e5295320fff42b4e6b3c3d6f2c88abc48d4b319290aca58392c560dc3b"} Dec 10 19:22:47 crc kubenswrapper[4828]: I1210 19:22:47.480869 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.745937948 podStartE2EDuration="7.480851318s" podCreationTimestamp="2025-12-10 19:22:40 +0000 UTC" firstStartedPulling="2025-12-10 19:22:42.087535581 +0000 UTC m=+1642.598146586" lastFinishedPulling="2025-12-10 19:22:46.822448951 +0000 UTC m=+1647.333059956" observedRunningTime="2025-12-10 19:22:47.480139499 +0000 UTC m=+1647.990750514" watchObservedRunningTime="2025-12-10 19:22:47.480851318 +0000 UTC m=+1647.991462323" Dec 10 19:22:48 crc kubenswrapper[4828]: I1210 19:22:48.797250 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:22:48 crc kubenswrapper[4828]: E1210 19:22:48.797925 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:22:48 crc kubenswrapper[4828]: I1210 19:22:48.919102 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-v9g52" Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.002271 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ztlt\" (UniqueName: \"kubernetes.io/projected/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-kube-api-access-5ztlt\") pod \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\" (UID: \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\") " Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.002368 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-combined-ca-bundle\") pod \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\" (UID: \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\") " Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.002462 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-config-data\") pod \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\" (UID: \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\") " Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.002569 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-scripts\") pod \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\" (UID: \"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f\") " Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.009048 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-scripts" (OuterVolumeSpecName: "scripts") pod "ac39304a-7ef4-42a7-a9b3-9ea15a237a4f" (UID: "ac39304a-7ef4-42a7-a9b3-9ea15a237a4f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.009895 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-kube-api-access-5ztlt" (OuterVolumeSpecName: "kube-api-access-5ztlt") pod "ac39304a-7ef4-42a7-a9b3-9ea15a237a4f" (UID: "ac39304a-7ef4-42a7-a9b3-9ea15a237a4f"). InnerVolumeSpecName "kube-api-access-5ztlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.036052 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-config-data" (OuterVolumeSpecName: "config-data") pod "ac39304a-7ef4-42a7-a9b3-9ea15a237a4f" (UID: "ac39304a-7ef4-42a7-a9b3-9ea15a237a4f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.045384 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac39304a-7ef4-42a7-a9b3-9ea15a237a4f" (UID: "ac39304a-7ef4-42a7-a9b3-9ea15a237a4f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.105638 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.105676 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.105685 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.105693 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ztlt\" (UniqueName: \"kubernetes.io/projected/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f-kube-api-access-5ztlt\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.482155 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-v9g52" event={"ID":"ac39304a-7ef4-42a7-a9b3-9ea15a237a4f","Type":"ContainerDied","Data":"74ad506bf9615e41e868725c2a8d2f29cb9c2e4510033adffa492c810e6c1f4e"} Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.482430 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74ad506bf9615e41e868725c2a8d2f29cb9c2e4510033adffa492c810e6c1f4e" Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.482210 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-v9g52" Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.528224 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.529241 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.534680 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.684330 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.684584 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f7c47155-fd88-4adb-962e-caf36152a6f8" containerName="nova-scheduler-scheduler" containerID="cri-o://4dbea57c3646137d7bc461d34f0d6c543f93a6f37639f833a0893e717b98d4d6" gracePeriod=30 Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.696333 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.696568 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="37407428-f199-4082-97ea-d9568eddfc1f" containerName="nova-api-log" containerID="cri-o://0f032ae026aa4205af0fa013963be337918313790e0b941250faba503a41d333" gracePeriod=30 Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.696698 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="37407428-f199-4082-97ea-d9568eddfc1f" containerName="nova-api-api" containerID="cri-o://854f52346883105f81eaf3a28670fdc042e5fa5f097525eeeb540d11f8d0c7ad" gracePeriod=30 Dec 10 19:22:49 crc kubenswrapper[4828]: I1210 19:22:49.716146 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:22:50 crc kubenswrapper[4828]: I1210 19:22:50.503657 4828 generic.go:334] "Generic (PLEG): container finished" podID="f7c47155-fd88-4adb-962e-caf36152a6f8" containerID="4dbea57c3646137d7bc461d34f0d6c543f93a6f37639f833a0893e717b98d4d6" exitCode=0 Dec 10 19:22:50 crc kubenswrapper[4828]: I1210 19:22:50.503752 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7c47155-fd88-4adb-962e-caf36152a6f8","Type":"ContainerDied","Data":"4dbea57c3646137d7bc461d34f0d6c543f93a6f37639f833a0893e717b98d4d6"} Dec 10 19:22:50 crc kubenswrapper[4828]: I1210 19:22:50.506420 4828 generic.go:334] "Generic (PLEG): container finished" podID="37407428-f199-4082-97ea-d9568eddfc1f" containerID="0f032ae026aa4205af0fa013963be337918313790e0b941250faba503a41d333" exitCode=143 Dec 10 19:22:50 crc kubenswrapper[4828]: I1210 19:22:50.507472 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37407428-f199-4082-97ea-d9568eddfc1f","Type":"ContainerDied","Data":"0f032ae026aa4205af0fa013963be337918313790e0b941250faba503a41d333"} Dec 10 19:22:50 crc kubenswrapper[4828]: I1210 19:22:50.513635 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 19:22:50 crc kubenswrapper[4828]: I1210 19:22:50.811011 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:22:50 crc kubenswrapper[4828]: I1210 19:22:50.966625 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldtrx\" (UniqueName: \"kubernetes.io/projected/f7c47155-fd88-4adb-962e-caf36152a6f8-kube-api-access-ldtrx\") pod \"f7c47155-fd88-4adb-962e-caf36152a6f8\" (UID: \"f7c47155-fd88-4adb-962e-caf36152a6f8\") " Dec 10 19:22:50 crc kubenswrapper[4828]: I1210 19:22:50.966773 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7c47155-fd88-4adb-962e-caf36152a6f8-combined-ca-bundle\") pod \"f7c47155-fd88-4adb-962e-caf36152a6f8\" (UID: \"f7c47155-fd88-4adb-962e-caf36152a6f8\") " Dec 10 19:22:50 crc kubenswrapper[4828]: I1210 19:22:50.966922 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7c47155-fd88-4adb-962e-caf36152a6f8-config-data\") pod \"f7c47155-fd88-4adb-962e-caf36152a6f8\" (UID: \"f7c47155-fd88-4adb-962e-caf36152a6f8\") " Dec 10 19:22:50 crc kubenswrapper[4828]: I1210 19:22:50.973133 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7c47155-fd88-4adb-962e-caf36152a6f8-kube-api-access-ldtrx" (OuterVolumeSpecName: "kube-api-access-ldtrx") pod "f7c47155-fd88-4adb-962e-caf36152a6f8" (UID: "f7c47155-fd88-4adb-962e-caf36152a6f8"). InnerVolumeSpecName "kube-api-access-ldtrx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:50 crc kubenswrapper[4828]: I1210 19:22:50.998561 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7c47155-fd88-4adb-962e-caf36152a6f8-config-data" (OuterVolumeSpecName: "config-data") pod "f7c47155-fd88-4adb-962e-caf36152a6f8" (UID: "f7c47155-fd88-4adb-962e-caf36152a6f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.008117 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7c47155-fd88-4adb-962e-caf36152a6f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7c47155-fd88-4adb-962e-caf36152a6f8" (UID: "f7c47155-fd88-4adb-962e-caf36152a6f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.070163 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7c47155-fd88-4adb-962e-caf36152a6f8-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.070225 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldtrx\" (UniqueName: \"kubernetes.io/projected/f7c47155-fd88-4adb-962e-caf36152a6f8-kube-api-access-ldtrx\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.070239 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7c47155-fd88-4adb-962e-caf36152a6f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.519246 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="db6cfeff-40e0-4bb8-ac4d-55cf6043537c" containerName="nova-metadata-log" containerID="cri-o://9f65f1571b18850e44d5356cad53d43aaa3286a9f054b406516f8f0bbdfff9a7" gracePeriod=30 Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.519860 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.523066 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7c47155-fd88-4adb-962e-caf36152a6f8","Type":"ContainerDied","Data":"c17bb023dfce3e592dcecbeb8b6cb1211c041959e1d7f18caf8c6a7d5d9d6d77"} Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.523101 4828 scope.go:117] "RemoveContainer" containerID="4dbea57c3646137d7bc461d34f0d6c543f93a6f37639f833a0893e717b98d4d6" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.523511 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="db6cfeff-40e0-4bb8-ac4d-55cf6043537c" containerName="nova-metadata-metadata" containerID="cri-o://9ac5a7c6345199c72ec2b239a21d08ad9a0bb7be487a20cdd00d11e9ba1f4e79" gracePeriod=30 Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.589854 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.601271 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.612651 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:22:51 crc kubenswrapper[4828]: E1210 19:22:51.613383 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c47155-fd88-4adb-962e-caf36152a6f8" containerName="nova-scheduler-scheduler" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.613410 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c47155-fd88-4adb-962e-caf36152a6f8" containerName="nova-scheduler-scheduler" Dec 10 19:22:51 crc kubenswrapper[4828]: E1210 19:22:51.613447 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac39304a-7ef4-42a7-a9b3-9ea15a237a4f" containerName="nova-manage" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.613457 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac39304a-7ef4-42a7-a9b3-9ea15a237a4f" containerName="nova-manage" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.613823 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac39304a-7ef4-42a7-a9b3-9ea15a237a4f" containerName="nova-manage" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.613854 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7c47155-fd88-4adb-962e-caf36152a6f8" containerName="nova-scheduler-scheduler" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.615637 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.618483 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.625548 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.688624 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f01fc94-86ea-4b2e-b89b-7279898f9366-config-data\") pod \"nova-scheduler-0\" (UID: \"1f01fc94-86ea-4b2e-b89b-7279898f9366\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.688955 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmd5j\" (UniqueName: \"kubernetes.io/projected/1f01fc94-86ea-4b2e-b89b-7279898f9366-kube-api-access-cmd5j\") pod \"nova-scheduler-0\" (UID: \"1f01fc94-86ea-4b2e-b89b-7279898f9366\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.689250 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f01fc94-86ea-4b2e-b89b-7279898f9366-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1f01fc94-86ea-4b2e-b89b-7279898f9366\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.791510 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f01fc94-86ea-4b2e-b89b-7279898f9366-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1f01fc94-86ea-4b2e-b89b-7279898f9366\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.791697 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f01fc94-86ea-4b2e-b89b-7279898f9366-config-data\") pod \"nova-scheduler-0\" (UID: \"1f01fc94-86ea-4b2e-b89b-7279898f9366\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.791768 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmd5j\" (UniqueName: \"kubernetes.io/projected/1f01fc94-86ea-4b2e-b89b-7279898f9366-kube-api-access-cmd5j\") pod \"nova-scheduler-0\" (UID: \"1f01fc94-86ea-4b2e-b89b-7279898f9366\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.796914 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f01fc94-86ea-4b2e-b89b-7279898f9366-config-data\") pod \"nova-scheduler-0\" (UID: \"1f01fc94-86ea-4b2e-b89b-7279898f9366\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.803424 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7c47155-fd88-4adb-962e-caf36152a6f8" path="/var/lib/kubelet/pods/f7c47155-fd88-4adb-962e-caf36152a6f8/volumes" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.809747 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmd5j\" (UniqueName: \"kubernetes.io/projected/1f01fc94-86ea-4b2e-b89b-7279898f9366-kube-api-access-cmd5j\") pod \"nova-scheduler-0\" (UID: \"1f01fc94-86ea-4b2e-b89b-7279898f9366\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.810137 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f01fc94-86ea-4b2e-b89b-7279898f9366-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1f01fc94-86ea-4b2e-b89b-7279898f9366\") " pod="openstack/nova-scheduler-0" Dec 10 19:22:51 crc kubenswrapper[4828]: I1210 19:22:51.989950 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 10 19:22:52 crc kubenswrapper[4828]: I1210 19:22:52.444669 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 10 19:22:52 crc kubenswrapper[4828]: W1210 19:22:52.446255 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f01fc94_86ea_4b2e_b89b_7279898f9366.slice/crio-1fb4fbc196f09dc52a789b28d5ff39eb0d06f8df0f9e0512a467001c4db36dca WatchSource:0}: Error finding container 1fb4fbc196f09dc52a789b28d5ff39eb0d06f8df0f9e0512a467001c4db36dca: Status 404 returned error can't find the container with id 1fb4fbc196f09dc52a789b28d5ff39eb0d06f8df0f9e0512a467001c4db36dca Dec 10 19:22:52 crc kubenswrapper[4828]: I1210 19:22:52.532426 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1f01fc94-86ea-4b2e-b89b-7279898f9366","Type":"ContainerStarted","Data":"1fb4fbc196f09dc52a789b28d5ff39eb0d06f8df0f9e0512a467001c4db36dca"} Dec 10 19:22:52 crc kubenswrapper[4828]: I1210 19:22:52.535159 4828 generic.go:334] "Generic (PLEG): container finished" podID="db6cfeff-40e0-4bb8-ac4d-55cf6043537c" containerID="9f65f1571b18850e44d5356cad53d43aaa3286a9f054b406516f8f0bbdfff9a7" exitCode=143 Dec 10 19:22:52 crc kubenswrapper[4828]: I1210 19:22:52.535226 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"db6cfeff-40e0-4bb8-ac4d-55cf6043537c","Type":"ContainerDied","Data":"9f65f1571b18850e44d5356cad53d43aaa3286a9f054b406516f8f0bbdfff9a7"} Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.306282 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.431842 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37407428-f199-4082-97ea-d9568eddfc1f-combined-ca-bundle\") pod \"37407428-f199-4082-97ea-d9568eddfc1f\" (UID: \"37407428-f199-4082-97ea-d9568eddfc1f\") " Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.431913 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37407428-f199-4082-97ea-d9568eddfc1f-logs\") pod \"37407428-f199-4082-97ea-d9568eddfc1f\" (UID: \"37407428-f199-4082-97ea-d9568eddfc1f\") " Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.431950 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gntv6\" (UniqueName: \"kubernetes.io/projected/37407428-f199-4082-97ea-d9568eddfc1f-kube-api-access-gntv6\") pod \"37407428-f199-4082-97ea-d9568eddfc1f\" (UID: \"37407428-f199-4082-97ea-d9568eddfc1f\") " Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.432226 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37407428-f199-4082-97ea-d9568eddfc1f-config-data\") pod \"37407428-f199-4082-97ea-d9568eddfc1f\" (UID: \"37407428-f199-4082-97ea-d9568eddfc1f\") " Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.434343 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37407428-f199-4082-97ea-d9568eddfc1f-logs" (OuterVolumeSpecName: "logs") pod "37407428-f199-4082-97ea-d9568eddfc1f" (UID: "37407428-f199-4082-97ea-d9568eddfc1f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.441681 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37407428-f199-4082-97ea-d9568eddfc1f-kube-api-access-gntv6" (OuterVolumeSpecName: "kube-api-access-gntv6") pod "37407428-f199-4082-97ea-d9568eddfc1f" (UID: "37407428-f199-4082-97ea-d9568eddfc1f"). InnerVolumeSpecName "kube-api-access-gntv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.466827 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37407428-f199-4082-97ea-d9568eddfc1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37407428-f199-4082-97ea-d9568eddfc1f" (UID: "37407428-f199-4082-97ea-d9568eddfc1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.467339 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37407428-f199-4082-97ea-d9568eddfc1f-config-data" (OuterVolumeSpecName: "config-data") pod "37407428-f199-4082-97ea-d9568eddfc1f" (UID: "37407428-f199-4082-97ea-d9568eddfc1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.535223 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37407428-f199-4082-97ea-d9568eddfc1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.535261 4828 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37407428-f199-4082-97ea-d9568eddfc1f-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.535276 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gntv6\" (UniqueName: \"kubernetes.io/projected/37407428-f199-4082-97ea-d9568eddfc1f-kube-api-access-gntv6\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.535291 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37407428-f199-4082-97ea-d9568eddfc1f-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.548886 4828 generic.go:334] "Generic (PLEG): container finished" podID="37407428-f199-4082-97ea-d9568eddfc1f" containerID="854f52346883105f81eaf3a28670fdc042e5fa5f097525eeeb540d11f8d0c7ad" exitCode=0 Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.548961 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37407428-f199-4082-97ea-d9568eddfc1f","Type":"ContainerDied","Data":"854f52346883105f81eaf3a28670fdc042e5fa5f097525eeeb540d11f8d0c7ad"} Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.549013 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37407428-f199-4082-97ea-d9568eddfc1f","Type":"ContainerDied","Data":"1774b820f2130c959e66a06790c9b36a94985cf3faab4185bda910c44af99ead"} Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.549038 4828 scope.go:117] "RemoveContainer" containerID="854f52346883105f81eaf3a28670fdc042e5fa5f097525eeeb540d11f8d0c7ad" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.549396 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.550721 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1f01fc94-86ea-4b2e-b89b-7279898f9366","Type":"ContainerStarted","Data":"321d9e244a332ae16447b3bb61433a7519563bf2259f3c36ca68a10c74428049"} Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.569119 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.56909812 podStartE2EDuration="2.56909812s" podCreationTimestamp="2025-12-10 19:22:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:22:53.568509734 +0000 UTC m=+1654.079120739" watchObservedRunningTime="2025-12-10 19:22:53.56909812 +0000 UTC m=+1654.079709125" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.597603 4828 scope.go:117] "RemoveContainer" containerID="0f032ae026aa4205af0fa013963be337918313790e0b941250faba503a41d333" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.601195 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.623500 4828 scope.go:117] "RemoveContainer" containerID="854f52346883105f81eaf3a28670fdc042e5fa5f097525eeeb540d11f8d0c7ad" Dec 10 19:22:53 crc kubenswrapper[4828]: E1210 19:22:53.626569 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"854f52346883105f81eaf3a28670fdc042e5fa5f097525eeeb540d11f8d0c7ad\": container with ID starting with 854f52346883105f81eaf3a28670fdc042e5fa5f097525eeeb540d11f8d0c7ad not found: ID does not exist" containerID="854f52346883105f81eaf3a28670fdc042e5fa5f097525eeeb540d11f8d0c7ad" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.626641 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"854f52346883105f81eaf3a28670fdc042e5fa5f097525eeeb540d11f8d0c7ad"} err="failed to get container status \"854f52346883105f81eaf3a28670fdc042e5fa5f097525eeeb540d11f8d0c7ad\": rpc error: code = NotFound desc = could not find container \"854f52346883105f81eaf3a28670fdc042e5fa5f097525eeeb540d11f8d0c7ad\": container with ID starting with 854f52346883105f81eaf3a28670fdc042e5fa5f097525eeeb540d11f8d0c7ad not found: ID does not exist" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.626673 4828 scope.go:117] "RemoveContainer" containerID="0f032ae026aa4205af0fa013963be337918313790e0b941250faba503a41d333" Dec 10 19:22:53 crc kubenswrapper[4828]: E1210 19:22:53.627306 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f032ae026aa4205af0fa013963be337918313790e0b941250faba503a41d333\": container with ID starting with 0f032ae026aa4205af0fa013963be337918313790e0b941250faba503a41d333 not found: ID does not exist" containerID="0f032ae026aa4205af0fa013963be337918313790e0b941250faba503a41d333" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.627411 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f032ae026aa4205af0fa013963be337918313790e0b941250faba503a41d333"} err="failed to get container status \"0f032ae026aa4205af0fa013963be337918313790e0b941250faba503a41d333\": rpc error: code = NotFound desc = could not find container \"0f032ae026aa4205af0fa013963be337918313790e0b941250faba503a41d333\": container with ID starting with 0f032ae026aa4205af0fa013963be337918313790e0b941250faba503a41d333 not found: ID does not exist" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.628632 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.640857 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 19:22:53 crc kubenswrapper[4828]: E1210 19:22:53.641513 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37407428-f199-4082-97ea-d9568eddfc1f" containerName="nova-api-api" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.641537 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="37407428-f199-4082-97ea-d9568eddfc1f" containerName="nova-api-api" Dec 10 19:22:53 crc kubenswrapper[4828]: E1210 19:22:53.641571 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37407428-f199-4082-97ea-d9568eddfc1f" containerName="nova-api-log" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.641581 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="37407428-f199-4082-97ea-d9568eddfc1f" containerName="nova-api-log" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.642028 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="37407428-f199-4082-97ea-d9568eddfc1f" containerName="nova-api-api" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.642059 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="37407428-f199-4082-97ea-d9568eddfc1f" containerName="nova-api-log" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.643785 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.646754 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.651905 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.739330 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50b1599c-e9bb-478e-aeb9-c311d903b466-config-data\") pod \"nova-api-0\" (UID: \"50b1599c-e9bb-478e-aeb9-c311d903b466\") " pod="openstack/nova-api-0" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.739625 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b1599c-e9bb-478e-aeb9-c311d903b466-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"50b1599c-e9bb-478e-aeb9-c311d903b466\") " pod="openstack/nova-api-0" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.739927 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdjgv\" (UniqueName: \"kubernetes.io/projected/50b1599c-e9bb-478e-aeb9-c311d903b466-kube-api-access-xdjgv\") pod \"nova-api-0\" (UID: \"50b1599c-e9bb-478e-aeb9-c311d903b466\") " pod="openstack/nova-api-0" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.739957 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50b1599c-e9bb-478e-aeb9-c311d903b466-logs\") pod \"nova-api-0\" (UID: \"50b1599c-e9bb-478e-aeb9-c311d903b466\") " pod="openstack/nova-api-0" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.803040 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37407428-f199-4082-97ea-d9568eddfc1f" path="/var/lib/kubelet/pods/37407428-f199-4082-97ea-d9568eddfc1f/volumes" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.842182 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdjgv\" (UniqueName: \"kubernetes.io/projected/50b1599c-e9bb-478e-aeb9-c311d903b466-kube-api-access-xdjgv\") pod \"nova-api-0\" (UID: \"50b1599c-e9bb-478e-aeb9-c311d903b466\") " pod="openstack/nova-api-0" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.842231 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50b1599c-e9bb-478e-aeb9-c311d903b466-logs\") pod \"nova-api-0\" (UID: \"50b1599c-e9bb-478e-aeb9-c311d903b466\") " pod="openstack/nova-api-0" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.842312 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50b1599c-e9bb-478e-aeb9-c311d903b466-config-data\") pod \"nova-api-0\" (UID: \"50b1599c-e9bb-478e-aeb9-c311d903b466\") " pod="openstack/nova-api-0" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.842434 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b1599c-e9bb-478e-aeb9-c311d903b466-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"50b1599c-e9bb-478e-aeb9-c311d903b466\") " pod="openstack/nova-api-0" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.843360 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50b1599c-e9bb-478e-aeb9-c311d903b466-logs\") pod \"nova-api-0\" (UID: \"50b1599c-e9bb-478e-aeb9-c311d903b466\") " pod="openstack/nova-api-0" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.846643 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b1599c-e9bb-478e-aeb9-c311d903b466-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"50b1599c-e9bb-478e-aeb9-c311d903b466\") " pod="openstack/nova-api-0" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.846704 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50b1599c-e9bb-478e-aeb9-c311d903b466-config-data\") pod \"nova-api-0\" (UID: \"50b1599c-e9bb-478e-aeb9-c311d903b466\") " pod="openstack/nova-api-0" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.858147 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdjgv\" (UniqueName: \"kubernetes.io/projected/50b1599c-e9bb-478e-aeb9-c311d903b466-kube-api-access-xdjgv\") pod \"nova-api-0\" (UID: \"50b1599c-e9bb-478e-aeb9-c311d903b466\") " pod="openstack/nova-api-0" Dec 10 19:22:53 crc kubenswrapper[4828]: I1210 19:22:53.964292 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:22:54 crc kubenswrapper[4828]: I1210 19:22:54.422048 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:22:54 crc kubenswrapper[4828]: I1210 19:22:54.588282 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"50b1599c-e9bb-478e-aeb9-c311d903b466","Type":"ContainerStarted","Data":"1ac9631451db9ba82709ab5f4038807cf0b52336f7ec3529f83b1edecc013dbe"} Dec 10 19:22:54 crc kubenswrapper[4828]: I1210 19:22:54.652852 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="db6cfeff-40e0-4bb8-ac4d-55cf6043537c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.245:8775/\": read tcp 10.217.0.2:52746->10.217.0.245:8775: read: connection reset by peer" Dec 10 19:22:54 crc kubenswrapper[4828]: I1210 19:22:54.652852 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="db6cfeff-40e0-4bb8-ac4d-55cf6043537c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.245:8775/\": read tcp 10.217.0.2:52732->10.217.0.245:8775: read: connection reset by peer" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.156207 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.280348 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-nova-metadata-tls-certs\") pod \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.280401 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-config-data\") pod \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.280553 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k88fd\" (UniqueName: \"kubernetes.io/projected/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-kube-api-access-k88fd\") pod \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.281716 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-combined-ca-bundle\") pod \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.281826 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-logs\") pod \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\" (UID: \"db6cfeff-40e0-4bb8-ac4d-55cf6043537c\") " Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.283149 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-logs" (OuterVolumeSpecName: "logs") pod "db6cfeff-40e0-4bb8-ac4d-55cf6043537c" (UID: "db6cfeff-40e0-4bb8-ac4d-55cf6043537c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.287526 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-kube-api-access-k88fd" (OuterVolumeSpecName: "kube-api-access-k88fd") pod "db6cfeff-40e0-4bb8-ac4d-55cf6043537c" (UID: "db6cfeff-40e0-4bb8-ac4d-55cf6043537c"). InnerVolumeSpecName "kube-api-access-k88fd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.317590 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db6cfeff-40e0-4bb8-ac4d-55cf6043537c" (UID: "db6cfeff-40e0-4bb8-ac4d-55cf6043537c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.321036 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-config-data" (OuterVolumeSpecName: "config-data") pod "db6cfeff-40e0-4bb8-ac4d-55cf6043537c" (UID: "db6cfeff-40e0-4bb8-ac4d-55cf6043537c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.354979 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "db6cfeff-40e0-4bb8-ac4d-55cf6043537c" (UID: "db6cfeff-40e0-4bb8-ac4d-55cf6043537c"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.385218 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.385258 4828 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.385268 4828 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.385279 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.385290 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k88fd\" (UniqueName: \"kubernetes.io/projected/db6cfeff-40e0-4bb8-ac4d-55cf6043537c-kube-api-access-k88fd\") on node \"crc\" DevicePath \"\"" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.605147 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"50b1599c-e9bb-478e-aeb9-c311d903b466","Type":"ContainerStarted","Data":"2b1e38ccb1a835a16df3418aaa3ddba896e5515e4bcd131135436b5ad05af363"} Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.605221 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"50b1599c-e9bb-478e-aeb9-c311d903b466","Type":"ContainerStarted","Data":"5459d74f5ea9fdeb46135081a7591d7b01e666182dfaf2c7de2d9af484a90e7e"} Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.607489 4828 generic.go:334] "Generic (PLEG): container finished" podID="db6cfeff-40e0-4bb8-ac4d-55cf6043537c" containerID="9ac5a7c6345199c72ec2b239a21d08ad9a0bb7be487a20cdd00d11e9ba1f4e79" exitCode=0 Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.607535 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"db6cfeff-40e0-4bb8-ac4d-55cf6043537c","Type":"ContainerDied","Data":"9ac5a7c6345199c72ec2b239a21d08ad9a0bb7be487a20cdd00d11e9ba1f4e79"} Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.607702 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"db6cfeff-40e0-4bb8-ac4d-55cf6043537c","Type":"ContainerDied","Data":"afbdbd00890f4f43a90831d8ca85136d4eb9e1d042d728573c62c25e19334f37"} Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.607562 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.607755 4828 scope.go:117] "RemoveContainer" containerID="9ac5a7c6345199c72ec2b239a21d08ad9a0bb7be487a20cdd00d11e9ba1f4e79" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.626297 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.6262791930000002 podStartE2EDuration="2.626279193s" podCreationTimestamp="2025-12-10 19:22:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:22:55.620600813 +0000 UTC m=+1656.131211818" watchObservedRunningTime="2025-12-10 19:22:55.626279193 +0000 UTC m=+1656.136890198" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.639581 4828 scope.go:117] "RemoveContainer" containerID="9f65f1571b18850e44d5356cad53d43aaa3286a9f054b406516f8f0bbdfff9a7" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.648996 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.660598 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.671170 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.671322 4828 scope.go:117] "RemoveContainer" containerID="9ac5a7c6345199c72ec2b239a21d08ad9a0bb7be487a20cdd00d11e9ba1f4e79" Dec 10 19:22:55 crc kubenswrapper[4828]: E1210 19:22:55.671871 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ac5a7c6345199c72ec2b239a21d08ad9a0bb7be487a20cdd00d11e9ba1f4e79\": container with ID starting with 9ac5a7c6345199c72ec2b239a21d08ad9a0bb7be487a20cdd00d11e9ba1f4e79 not found: ID does not exist" containerID="9ac5a7c6345199c72ec2b239a21d08ad9a0bb7be487a20cdd00d11e9ba1f4e79" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.671917 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ac5a7c6345199c72ec2b239a21d08ad9a0bb7be487a20cdd00d11e9ba1f4e79"} err="failed to get container status \"9ac5a7c6345199c72ec2b239a21d08ad9a0bb7be487a20cdd00d11e9ba1f4e79\": rpc error: code = NotFound desc = could not find container \"9ac5a7c6345199c72ec2b239a21d08ad9a0bb7be487a20cdd00d11e9ba1f4e79\": container with ID starting with 9ac5a7c6345199c72ec2b239a21d08ad9a0bb7be487a20cdd00d11e9ba1f4e79 not found: ID does not exist" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.671950 4828 scope.go:117] "RemoveContainer" containerID="9f65f1571b18850e44d5356cad53d43aaa3286a9f054b406516f8f0bbdfff9a7" Dec 10 19:22:55 crc kubenswrapper[4828]: E1210 19:22:55.672298 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db6cfeff-40e0-4bb8-ac4d-55cf6043537c" containerName="nova-metadata-log" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.672398 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="db6cfeff-40e0-4bb8-ac4d-55cf6043537c" containerName="nova-metadata-log" Dec 10 19:22:55 crc kubenswrapper[4828]: E1210 19:22:55.672478 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db6cfeff-40e0-4bb8-ac4d-55cf6043537c" containerName="nova-metadata-metadata" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.672587 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="db6cfeff-40e0-4bb8-ac4d-55cf6043537c" containerName="nova-metadata-metadata" Dec 10 19:22:55 crc kubenswrapper[4828]: E1210 19:22:55.672305 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f65f1571b18850e44d5356cad53d43aaa3286a9f054b406516f8f0bbdfff9a7\": container with ID starting with 9f65f1571b18850e44d5356cad53d43aaa3286a9f054b406516f8f0bbdfff9a7 not found: ID does not exist" containerID="9f65f1571b18850e44d5356cad53d43aaa3286a9f054b406516f8f0bbdfff9a7" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.672823 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f65f1571b18850e44d5356cad53d43aaa3286a9f054b406516f8f0bbdfff9a7"} err="failed to get container status \"9f65f1571b18850e44d5356cad53d43aaa3286a9f054b406516f8f0bbdfff9a7\": rpc error: code = NotFound desc = could not find container \"9f65f1571b18850e44d5356cad53d43aaa3286a9f054b406516f8f0bbdfff9a7\": container with ID starting with 9f65f1571b18850e44d5356cad53d43aaa3286a9f054b406516f8f0bbdfff9a7 not found: ID does not exist" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.673241 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="db6cfeff-40e0-4bb8-ac4d-55cf6043537c" containerName="nova-metadata-log" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.673343 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="db6cfeff-40e0-4bb8-ac4d-55cf6043537c" containerName="nova-metadata-metadata" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.675068 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.677949 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.678727 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.698617 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.799460 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98c456a-bb60-4ef8-8367-fad750d64487-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b98c456a-bb60-4ef8-8367-fad750d64487\") " pod="openstack/nova-metadata-0" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.799533 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b98c456a-bb60-4ef8-8367-fad750d64487-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b98c456a-bb60-4ef8-8367-fad750d64487\") " pod="openstack/nova-metadata-0" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.799894 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98c456a-bb60-4ef8-8367-fad750d64487-logs\") pod \"nova-metadata-0\" (UID: \"b98c456a-bb60-4ef8-8367-fad750d64487\") " pod="openstack/nova-metadata-0" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.800094 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98c456a-bb60-4ef8-8367-fad750d64487-config-data\") pod \"nova-metadata-0\" (UID: \"b98c456a-bb60-4ef8-8367-fad750d64487\") " pod="openstack/nova-metadata-0" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.800215 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp2zq\" (UniqueName: \"kubernetes.io/projected/b98c456a-bb60-4ef8-8367-fad750d64487-kube-api-access-wp2zq\") pod \"nova-metadata-0\" (UID: \"b98c456a-bb60-4ef8-8367-fad750d64487\") " pod="openstack/nova-metadata-0" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.800998 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db6cfeff-40e0-4bb8-ac4d-55cf6043537c" path="/var/lib/kubelet/pods/db6cfeff-40e0-4bb8-ac4d-55cf6043537c/volumes" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.901852 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98c456a-bb60-4ef8-8367-fad750d64487-logs\") pod \"nova-metadata-0\" (UID: \"b98c456a-bb60-4ef8-8367-fad750d64487\") " pod="openstack/nova-metadata-0" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.901979 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98c456a-bb60-4ef8-8367-fad750d64487-config-data\") pod \"nova-metadata-0\" (UID: \"b98c456a-bb60-4ef8-8367-fad750d64487\") " pod="openstack/nova-metadata-0" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.902047 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp2zq\" (UniqueName: \"kubernetes.io/projected/b98c456a-bb60-4ef8-8367-fad750d64487-kube-api-access-wp2zq\") pod \"nova-metadata-0\" (UID: \"b98c456a-bb60-4ef8-8367-fad750d64487\") " pod="openstack/nova-metadata-0" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.902109 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98c456a-bb60-4ef8-8367-fad750d64487-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b98c456a-bb60-4ef8-8367-fad750d64487\") " pod="openstack/nova-metadata-0" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.902144 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b98c456a-bb60-4ef8-8367-fad750d64487-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b98c456a-bb60-4ef8-8367-fad750d64487\") " pod="openstack/nova-metadata-0" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.902265 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b98c456a-bb60-4ef8-8367-fad750d64487-logs\") pod \"nova-metadata-0\" (UID: \"b98c456a-bb60-4ef8-8367-fad750d64487\") " pod="openstack/nova-metadata-0" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.905709 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b98c456a-bb60-4ef8-8367-fad750d64487-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b98c456a-bb60-4ef8-8367-fad750d64487\") " pod="openstack/nova-metadata-0" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.905762 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b98c456a-bb60-4ef8-8367-fad750d64487-config-data\") pod \"nova-metadata-0\" (UID: \"b98c456a-bb60-4ef8-8367-fad750d64487\") " pod="openstack/nova-metadata-0" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.905879 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b98c456a-bb60-4ef8-8367-fad750d64487-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b98c456a-bb60-4ef8-8367-fad750d64487\") " pod="openstack/nova-metadata-0" Dec 10 19:22:55 crc kubenswrapper[4828]: I1210 19:22:55.918250 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp2zq\" (UniqueName: \"kubernetes.io/projected/b98c456a-bb60-4ef8-8367-fad750d64487-kube-api-access-wp2zq\") pod \"nova-metadata-0\" (UID: \"b98c456a-bb60-4ef8-8367-fad750d64487\") " pod="openstack/nova-metadata-0" Dec 10 19:22:56 crc kubenswrapper[4828]: I1210 19:22:56.000747 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 10 19:22:56 crc kubenswrapper[4828]: I1210 19:22:56.449421 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 10 19:22:56 crc kubenswrapper[4828]: W1210 19:22:56.450564 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb98c456a_bb60_4ef8_8367_fad750d64487.slice/crio-c67266b3a2a06bde651daa41165eda484d36f2d285e26fd36f2ae8dfb8b1b5ff WatchSource:0}: Error finding container c67266b3a2a06bde651daa41165eda484d36f2d285e26fd36f2ae8dfb8b1b5ff: Status 404 returned error can't find the container with id c67266b3a2a06bde651daa41165eda484d36f2d285e26fd36f2ae8dfb8b1b5ff Dec 10 19:22:56 crc kubenswrapper[4828]: I1210 19:22:56.619977 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b98c456a-bb60-4ef8-8367-fad750d64487","Type":"ContainerStarted","Data":"c67266b3a2a06bde651daa41165eda484d36f2d285e26fd36f2ae8dfb8b1b5ff"} Dec 10 19:22:56 crc kubenswrapper[4828]: I1210 19:22:56.990772 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 10 19:22:57 crc kubenswrapper[4828]: I1210 19:22:57.640459 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b98c456a-bb60-4ef8-8367-fad750d64487","Type":"ContainerStarted","Data":"ab5d6d2eefc5c4d99d4ac1500ad444bd09398cf28325f8f04deb65ef910e24f3"} Dec 10 19:22:57 crc kubenswrapper[4828]: I1210 19:22:57.640519 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b98c456a-bb60-4ef8-8367-fad750d64487","Type":"ContainerStarted","Data":"d4b02bda887cb964e6ab94ada73664ac8e3613a376d30860ecf0cff1003f0934"} Dec 10 19:22:57 crc kubenswrapper[4828]: I1210 19:22:57.671096 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.671070657 podStartE2EDuration="2.671070657s" podCreationTimestamp="2025-12-10 19:22:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:22:57.659572293 +0000 UTC m=+1658.170183318" watchObservedRunningTime="2025-12-10 19:22:57.671070657 +0000 UTC m=+1658.181681662" Dec 10 19:23:00 crc kubenswrapper[4828]: I1210 19:23:00.789315 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:23:00 crc kubenswrapper[4828]: E1210 19:23:00.790194 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:23:01 crc kubenswrapper[4828]: I1210 19:23:01.001567 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 19:23:01 crc kubenswrapper[4828]: I1210 19:23:01.001935 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 10 19:23:01 crc kubenswrapper[4828]: I1210 19:23:01.991073 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 10 19:23:02 crc kubenswrapper[4828]: I1210 19:23:02.020582 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 10 19:23:02 crc kubenswrapper[4828]: I1210 19:23:02.717689 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 10 19:23:03 crc kubenswrapper[4828]: I1210 19:23:03.965202 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 19:23:03 crc kubenswrapper[4828]: I1210 19:23:03.965600 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 19:23:05 crc kubenswrapper[4828]: I1210 19:23:05.047020 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="50b1599c-e9bb-478e-aeb9-c311d903b466" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.252:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 19:23:05 crc kubenswrapper[4828]: I1210 19:23:05.047034 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="50b1599c-e9bb-478e-aeb9-c311d903b466" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.252:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 19:23:06 crc kubenswrapper[4828]: I1210 19:23:06.001680 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 19:23:06 crc kubenswrapper[4828]: I1210 19:23:06.002945 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 10 19:23:07 crc kubenswrapper[4828]: I1210 19:23:07.014054 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b98c456a-bb60-4ef8-8367-fad750d64487" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.253:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 19:23:07 crc kubenswrapper[4828]: I1210 19:23:07.014042 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b98c456a-bb60-4ef8-8367-fad750d64487" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.253:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 19:23:10 crc kubenswrapper[4828]: I1210 19:23:10.805677 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 10 19:23:12 crc kubenswrapper[4828]: I1210 19:23:12.790508 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:23:12 crc kubenswrapper[4828]: E1210 19:23:12.790983 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:23:12 crc kubenswrapper[4828]: I1210 19:23:12.803895 4828 generic.go:334] "Generic (PLEG): container finished" podID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerID="bc200df36df7c83b3e98bfd8302854d7b057f40b4b3c3aff31ae44ecb938792a" exitCode=137 Dec 10 19:23:12 crc kubenswrapper[4828]: I1210 19:23:12.803941 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"60ef7c6c-dfbd-4fff-b451-80b21458f86b","Type":"ContainerDied","Data":"bc200df36df7c83b3e98bfd8302854d7b057f40b4b3c3aff31ae44ecb938792a"} Dec 10 19:23:12 crc kubenswrapper[4828]: I1210 19:23:12.803973 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"60ef7c6c-dfbd-4fff-b451-80b21458f86b","Type":"ContainerDied","Data":"da1b36078c5882c12ba56845d3e3c12b90ca993e6e46b564a88c2ab153a7a1c7"} Dec 10 19:23:12 crc kubenswrapper[4828]: I1210 19:23:12.803986 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da1b36078c5882c12ba56845d3e3c12b90ca993e6e46b564a88c2ab153a7a1c7" Dec 10 19:23:12 crc kubenswrapper[4828]: I1210 19:23:12.831395 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 10 19:23:12 crc kubenswrapper[4828]: I1210 19:23:12.998670 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-scripts\") pod \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\" (UID: \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\") " Dec 10 19:23:12 crc kubenswrapper[4828]: I1210 19:23:12.998755 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-combined-ca-bundle\") pod \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\" (UID: \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\") " Dec 10 19:23:12 crc kubenswrapper[4828]: I1210 19:23:12.998948 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfpw4\" (UniqueName: \"kubernetes.io/projected/60ef7c6c-dfbd-4fff-b451-80b21458f86b-kube-api-access-jfpw4\") pod \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\" (UID: \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\") " Dec 10 19:23:12 crc kubenswrapper[4828]: I1210 19:23:12.999120 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-config-data\") pod \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\" (UID: \"60ef7c6c-dfbd-4fff-b451-80b21458f86b\") " Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.029968 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-scripts" (OuterVolumeSpecName: "scripts") pod "60ef7c6c-dfbd-4fff-b451-80b21458f86b" (UID: "60ef7c6c-dfbd-4fff-b451-80b21458f86b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.033083 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60ef7c6c-dfbd-4fff-b451-80b21458f86b-kube-api-access-jfpw4" (OuterVolumeSpecName: "kube-api-access-jfpw4") pod "60ef7c6c-dfbd-4fff-b451-80b21458f86b" (UID: "60ef7c6c-dfbd-4fff-b451-80b21458f86b"). InnerVolumeSpecName "kube-api-access-jfpw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.101995 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.102032 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfpw4\" (UniqueName: \"kubernetes.io/projected/60ef7c6c-dfbd-4fff-b451-80b21458f86b-kube-api-access-jfpw4\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.201308 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-config-data" (OuterVolumeSpecName: "config-data") pod "60ef7c6c-dfbd-4fff-b451-80b21458f86b" (UID: "60ef7c6c-dfbd-4fff-b451-80b21458f86b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.203622 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.230457 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60ef7c6c-dfbd-4fff-b451-80b21458f86b" (UID: "60ef7c6c-dfbd-4fff-b451-80b21458f86b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.305380 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60ef7c6c-dfbd-4fff-b451-80b21458f86b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.812912 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.842388 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.858007 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.870934 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 10 19:23:13 crc kubenswrapper[4828]: E1210 19:23:13.871505 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerName="aodh-notifier" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.871527 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerName="aodh-notifier" Dec 10 19:23:13 crc kubenswrapper[4828]: E1210 19:23:13.871587 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerName="aodh-listener" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.871598 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerName="aodh-listener" Dec 10 19:23:13 crc kubenswrapper[4828]: E1210 19:23:13.871618 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerName="aodh-api" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.871626 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerName="aodh-api" Dec 10 19:23:13 crc kubenswrapper[4828]: E1210 19:23:13.871642 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerName="aodh-evaluator" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.871650 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerName="aodh-evaluator" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.871969 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerName="aodh-notifier" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.872005 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerName="aodh-api" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.872032 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerName="aodh-listener" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.872053 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" containerName="aodh-evaluator" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.874231 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.881040 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.881683 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.881738 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.881995 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.883565 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.883965 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-qg8rx" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.916999 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-internal-tls-certs\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.917262 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-public-tls-certs\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.917330 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-config-data\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.917730 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xptc\" (UniqueName: \"kubernetes.io/projected/c3ea5ad2-3e85-405a-8e36-01578676f4e9-kube-api-access-4xptc\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.917872 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-scripts\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.917897 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-combined-ca-bundle\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.968011 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.968452 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.968563 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 19:23:13 crc kubenswrapper[4828]: I1210 19:23:13.973827 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 19:23:14 crc kubenswrapper[4828]: I1210 19:23:14.020938 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xptc\" (UniqueName: \"kubernetes.io/projected/c3ea5ad2-3e85-405a-8e36-01578676f4e9-kube-api-access-4xptc\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:14 crc kubenswrapper[4828]: I1210 19:23:14.021036 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-scripts\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:14 crc kubenswrapper[4828]: I1210 19:23:14.021061 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-combined-ca-bundle\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:14 crc kubenswrapper[4828]: I1210 19:23:14.021175 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-internal-tls-certs\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:14 crc kubenswrapper[4828]: I1210 19:23:14.021273 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-public-tls-certs\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:14 crc kubenswrapper[4828]: I1210 19:23:14.021310 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-config-data\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:14 crc kubenswrapper[4828]: I1210 19:23:14.027185 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-scripts\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:14 crc kubenswrapper[4828]: I1210 19:23:14.028115 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-config-data\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:14 crc kubenswrapper[4828]: I1210 19:23:14.028898 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-combined-ca-bundle\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:14 crc kubenswrapper[4828]: I1210 19:23:14.029492 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-internal-tls-certs\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:14 crc kubenswrapper[4828]: I1210 19:23:14.032294 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-public-tls-certs\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:14 crc kubenswrapper[4828]: I1210 19:23:14.041272 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xptc\" (UniqueName: \"kubernetes.io/projected/c3ea5ad2-3e85-405a-8e36-01578676f4e9-kube-api-access-4xptc\") pod \"aodh-0\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " pod="openstack/aodh-0" Dec 10 19:23:14 crc kubenswrapper[4828]: I1210 19:23:14.194033 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 10 19:23:14 crc kubenswrapper[4828]: W1210 19:23:14.679872 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3ea5ad2_3e85_405a_8e36_01578676f4e9.slice/crio-24cc3143556092f8e7df8f375a3fad56a976d0cd66ad6fa849e2888b713df707 WatchSource:0}: Error finding container 24cc3143556092f8e7df8f375a3fad56a976d0cd66ad6fa849e2888b713df707: Status 404 returned error can't find the container with id 24cc3143556092f8e7df8f375a3fad56a976d0cd66ad6fa849e2888b713df707 Dec 10 19:23:14 crc kubenswrapper[4828]: I1210 19:23:14.682891 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 10 19:23:14 crc kubenswrapper[4828]: I1210 19:23:14.824419 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c3ea5ad2-3e85-405a-8e36-01578676f4e9","Type":"ContainerStarted","Data":"24cc3143556092f8e7df8f375a3fad56a976d0cd66ad6fa849e2888b713df707"} Dec 10 19:23:14 crc kubenswrapper[4828]: I1210 19:23:14.824639 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 19:23:14 crc kubenswrapper[4828]: I1210 19:23:14.828153 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.029421 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-ncgpx"] Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.032378 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.062285 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-ncgpx"] Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.160024 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7nqd\" (UniqueName: \"kubernetes.io/projected/0c16730a-9ba1-48c7-93f8-f10980ca44ab-kube-api-access-w7nqd\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.160089 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.160179 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.160204 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.160250 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-config\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.160516 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.263051 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7nqd\" (UniqueName: \"kubernetes.io/projected/0c16730a-9ba1-48c7-93f8-f10980ca44ab-kube-api-access-w7nqd\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.263126 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.263222 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.263251 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.263298 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-config\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.263393 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.264471 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.264541 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-config\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.264598 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.264706 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.264915 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.282862 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7nqd\" (UniqueName: \"kubernetes.io/projected/0c16730a-9ba1-48c7-93f8-f10980ca44ab-kube-api-access-w7nqd\") pod \"dnsmasq-dns-f84f9ccf-ncgpx\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.384311 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.803692 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60ef7c6c-dfbd-4fff-b451-80b21458f86b" path="/var/lib/kubelet/pods/60ef7c6c-dfbd-4fff-b451-80b21458f86b/volumes" Dec 10 19:23:15 crc kubenswrapper[4828]: W1210 19:23:15.929298 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c16730a_9ba1_48c7_93f8_f10980ca44ab.slice/crio-6adaf9f1012dc20dcceac3b0234eab63f3c3bcd8a0c8c6018903402a3a80b7cf WatchSource:0}: Error finding container 6adaf9f1012dc20dcceac3b0234eab63f3c3bcd8a0c8c6018903402a3a80b7cf: Status 404 returned error can't find the container with id 6adaf9f1012dc20dcceac3b0234eab63f3c3bcd8a0c8c6018903402a3a80b7cf Dec 10 19:23:15 crc kubenswrapper[4828]: I1210 19:23:15.931406 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-ncgpx"] Dec 10 19:23:16 crc kubenswrapper[4828]: I1210 19:23:16.008824 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 19:23:16 crc kubenswrapper[4828]: I1210 19:23:16.021371 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 10 19:23:16 crc kubenswrapper[4828]: I1210 19:23:16.035522 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 19:23:16 crc kubenswrapper[4828]: I1210 19:23:16.851262 4828 generic.go:334] "Generic (PLEG): container finished" podID="0c16730a-9ba1-48c7-93f8-f10980ca44ab" containerID="6b91b4ce91d9e2ad35618daa9cc5cf8f27497a876bd80f937b05a985356835ad" exitCode=0 Dec 10 19:23:16 crc kubenswrapper[4828]: I1210 19:23:16.851347 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" event={"ID":"0c16730a-9ba1-48c7-93f8-f10980ca44ab","Type":"ContainerDied","Data":"6b91b4ce91d9e2ad35618daa9cc5cf8f27497a876bd80f937b05a985356835ad"} Dec 10 19:23:16 crc kubenswrapper[4828]: I1210 19:23:16.851872 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" event={"ID":"0c16730a-9ba1-48c7-93f8-f10980ca44ab","Type":"ContainerStarted","Data":"6adaf9f1012dc20dcceac3b0234eab63f3c3bcd8a0c8c6018903402a3a80b7cf"} Dec 10 19:23:16 crc kubenswrapper[4828]: I1210 19:23:16.857324 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c3ea5ad2-3e85-405a-8e36-01578676f4e9","Type":"ContainerStarted","Data":"82c6184925fd5fc91a8587873d8fb36c3832969f942270b28d1e662c0ac33dd9"} Dec 10 19:23:16 crc kubenswrapper[4828]: I1210 19:23:16.863916 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.349780 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.350429 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerName="ceilometer-central-agent" containerID="cri-o://46bb815020c90cd623232c5f1e70bc56377d205a168df279124bb5cacf240003" gracePeriod=30 Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.350493 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerName="proxy-httpd" containerID="cri-o://ed4bde243a6be79ae2b68d87435d741c6a9953ed37a775ae40496dbf028d88ac" gracePeriod=30 Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.350542 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerName="sg-core" containerID="cri-o://757b304b97d21f6ae4063f6a0334b634076c035c88a67045f183b3997fa16c51" gracePeriod=30 Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.350586 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerName="ceilometer-notification-agent" containerID="cri-o://332a9ba66bc89a6c7a57b8052cb0fe701f866bb29da3b489d148e7c37821e1bb" gracePeriod=30 Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.398312 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.398546 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="7cd541f6-4cd3-42b6-9d59-ad4b6246044e" containerName="kube-state-metrics" containerID="cri-o://0e61cbc0b1b55deb42c336115cad8dfaa9129689804f341614c3edf3541b1cda" gracePeriod=30 Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.494616 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.494895 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mysqld-exporter-0" podUID="9fd3d999-4b0d-422c-a7ae-01065d577c38" containerName="mysqld-exporter" containerID="cri-o://2a55dae265e67734536309d37d91f9df1cc16e0fbe174d434f83c5f5bda21d5f" gracePeriod=30 Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.882304 4828 generic.go:334] "Generic (PLEG): container finished" podID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerID="ed4bde243a6be79ae2b68d87435d741c6a9953ed37a775ae40496dbf028d88ac" exitCode=0 Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.882346 4828 generic.go:334] "Generic (PLEG): container finished" podID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerID="757b304b97d21f6ae4063f6a0334b634076c035c88a67045f183b3997fa16c51" exitCode=2 Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.882357 4828 generic.go:334] "Generic (PLEG): container finished" podID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerID="46bb815020c90cd623232c5f1e70bc56377d205a168df279124bb5cacf240003" exitCode=0 Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.882380 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9cbf2a7-b956-4bf8-90f9-1987aafe08df","Type":"ContainerDied","Data":"ed4bde243a6be79ae2b68d87435d741c6a9953ed37a775ae40496dbf028d88ac"} Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.882425 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9cbf2a7-b956-4bf8-90f9-1987aafe08df","Type":"ContainerDied","Data":"757b304b97d21f6ae4063f6a0334b634076c035c88a67045f183b3997fa16c51"} Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.882437 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9cbf2a7-b956-4bf8-90f9-1987aafe08df","Type":"ContainerDied","Data":"46bb815020c90cd623232c5f1e70bc56377d205a168df279124bb5cacf240003"} Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.887125 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" event={"ID":"0c16730a-9ba1-48c7-93f8-f10980ca44ab","Type":"ContainerStarted","Data":"7cceaf33cfacf6372e9726e3cbed0ef4b721b062ce3a3307e4305f4313c61666"} Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.887574 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.901578 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c3ea5ad2-3e85-405a-8e36-01578676f4e9","Type":"ContainerStarted","Data":"bd19855d1153a299acb9904c79a37fc78b3e4a29e14b371d6a11db2bfff5989a"} Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.912283 4828 generic.go:334] "Generic (PLEG): container finished" podID="9fd3d999-4b0d-422c-a7ae-01065d577c38" containerID="2a55dae265e67734536309d37d91f9df1cc16e0fbe174d434f83c5f5bda21d5f" exitCode=2 Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.912387 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"9fd3d999-4b0d-422c-a7ae-01065d577c38","Type":"ContainerDied","Data":"2a55dae265e67734536309d37d91f9df1cc16e0fbe174d434f83c5f5bda21d5f"} Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.922112 4828 generic.go:334] "Generic (PLEG): container finished" podID="7cd541f6-4cd3-42b6-9d59-ad4b6246044e" containerID="0e61cbc0b1b55deb42c336115cad8dfaa9129689804f341614c3edf3541b1cda" exitCode=2 Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.922361 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7cd541f6-4cd3-42b6-9d59-ad4b6246044e","Type":"ContainerDied","Data":"0e61cbc0b1b55deb42c336115cad8dfaa9129689804f341614c3edf3541b1cda"} Dec 10 19:23:17 crc kubenswrapper[4828]: I1210 19:23:17.937290 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" podStartSLOduration=3.937270822 podStartE2EDuration="3.937270822s" podCreationTimestamp="2025-12-10 19:23:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:23:17.909860436 +0000 UTC m=+1678.420471441" watchObservedRunningTime="2025-12-10 19:23:17.937270822 +0000 UTC m=+1678.447881827" Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.080298 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.139676 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csdml\" (UniqueName: \"kubernetes.io/projected/7cd541f6-4cd3-42b6-9d59-ad4b6246044e-kube-api-access-csdml\") pod \"7cd541f6-4cd3-42b6-9d59-ad4b6246044e\" (UID: \"7cd541f6-4cd3-42b6-9d59-ad4b6246044e\") " Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.148304 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cd541f6-4cd3-42b6-9d59-ad4b6246044e-kube-api-access-csdml" (OuterVolumeSpecName: "kube-api-access-csdml") pod "7cd541f6-4cd3-42b6-9d59-ad4b6246044e" (UID: "7cd541f6-4cd3-42b6-9d59-ad4b6246044e"). InnerVolumeSpecName "kube-api-access-csdml". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.222647 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.243902 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csdml\" (UniqueName: \"kubernetes.io/projected/7cd541f6-4cd3-42b6-9d59-ad4b6246044e-kube-api-access-csdml\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.345391 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fd3d999-4b0d-422c-a7ae-01065d577c38-config-data\") pod \"9fd3d999-4b0d-422c-a7ae-01065d577c38\" (UID: \"9fd3d999-4b0d-422c-a7ae-01065d577c38\") " Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.345694 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd3d999-4b0d-422c-a7ae-01065d577c38-combined-ca-bundle\") pod \"9fd3d999-4b0d-422c-a7ae-01065d577c38\" (UID: \"9fd3d999-4b0d-422c-a7ae-01065d577c38\") " Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.345874 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfw5c\" (UniqueName: \"kubernetes.io/projected/9fd3d999-4b0d-422c-a7ae-01065d577c38-kube-api-access-rfw5c\") pod \"9fd3d999-4b0d-422c-a7ae-01065d577c38\" (UID: \"9fd3d999-4b0d-422c-a7ae-01065d577c38\") " Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.352947 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fd3d999-4b0d-422c-a7ae-01065d577c38-kube-api-access-rfw5c" (OuterVolumeSpecName: "kube-api-access-rfw5c") pod "9fd3d999-4b0d-422c-a7ae-01065d577c38" (UID: "9fd3d999-4b0d-422c-a7ae-01065d577c38"). InnerVolumeSpecName "kube-api-access-rfw5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.395976 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd3d999-4b0d-422c-a7ae-01065d577c38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9fd3d999-4b0d-422c-a7ae-01065d577c38" (UID: "9fd3d999-4b0d-422c-a7ae-01065d577c38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.423219 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd3d999-4b0d-422c-a7ae-01065d577c38-config-data" (OuterVolumeSpecName: "config-data") pod "9fd3d999-4b0d-422c-a7ae-01065d577c38" (UID: "9fd3d999-4b0d-422c-a7ae-01065d577c38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.448756 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd3d999-4b0d-422c-a7ae-01065d577c38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.449016 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfw5c\" (UniqueName: \"kubernetes.io/projected/9fd3d999-4b0d-422c-a7ae-01065d577c38-kube-api-access-rfw5c\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.449136 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fd3d999-4b0d-422c-a7ae-01065d577c38-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.935373 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.936311 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7cd541f6-4cd3-42b6-9d59-ad4b6246044e","Type":"ContainerDied","Data":"31d7712b7667da8678af7c12d0e60c79cc9d8779cfd4545f4f0adb5aeb896450"} Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.936374 4828 scope.go:117] "RemoveContainer" containerID="0e61cbc0b1b55deb42c336115cad8dfaa9129689804f341614c3edf3541b1cda" Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.939951 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c3ea5ad2-3e85-405a-8e36-01578676f4e9","Type":"ContainerStarted","Data":"db3ae826961c858928c3278e11bbc8cf3855efc81b2c03480a7363089c12f59d"} Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.939998 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c3ea5ad2-3e85-405a-8e36-01578676f4e9","Type":"ContainerStarted","Data":"5a817e7054f0f1e9d6f8499822b0b209fe69c5141846f5cd362dd2014e5dfeae"} Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.951600 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.951634 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"9fd3d999-4b0d-422c-a7ae-01065d577c38","Type":"ContainerDied","Data":"546859861b4dd77704052864b69c1da3edffa0082c086146dca7634985726909"} Dec 10 19:23:18 crc kubenswrapper[4828]: I1210 19:23:18.993746 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.188892813 podStartE2EDuration="5.99372147s" podCreationTimestamp="2025-12-10 19:23:13 +0000 UTC" firstStartedPulling="2025-12-10 19:23:14.682263006 +0000 UTC m=+1675.192874011" lastFinishedPulling="2025-12-10 19:23:18.487091663 +0000 UTC m=+1678.997702668" observedRunningTime="2025-12-10 19:23:18.964536217 +0000 UTC m=+1679.475147212" watchObservedRunningTime="2025-12-10 19:23:18.99372147 +0000 UTC m=+1679.504332465" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.041664 4828 scope.go:117] "RemoveContainer" containerID="2a55dae265e67734536309d37d91f9df1cc16e0fbe174d434f83c5f5bda21d5f" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.104692 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.146859 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.170227 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.206274 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.230858 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Dec 10 19:23:19 crc kubenswrapper[4828]: E1210 19:23:19.231442 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cd541f6-4cd3-42b6-9d59-ad4b6246044e" containerName="kube-state-metrics" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.231467 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cd541f6-4cd3-42b6-9d59-ad4b6246044e" containerName="kube-state-metrics" Dec 10 19:23:19 crc kubenswrapper[4828]: E1210 19:23:19.231497 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fd3d999-4b0d-422c-a7ae-01065d577c38" containerName="mysqld-exporter" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.231505 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fd3d999-4b0d-422c-a7ae-01065d577c38" containerName="mysqld-exporter" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.231706 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cd541f6-4cd3-42b6-9d59-ad4b6246044e" containerName="kube-state-metrics" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.231736 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fd3d999-4b0d-422c-a7ae-01065d577c38" containerName="mysqld-exporter" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.232725 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.240238 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-mysqld-exporter-svc" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.240863 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.255186 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.277956 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.280737 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.293294 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.293358 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.294752 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbkpr\" (UniqueName: \"kubernetes.io/projected/198e58a9-696b-4857-a6fe-f31cf31d61d7-kube-api-access-cbkpr\") pod \"mysqld-exporter-0\" (UID: \"198e58a9-696b-4857-a6fe-f31cf31d61d7\") " pod="openstack/mysqld-exporter-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.294839 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/198e58a9-696b-4857-a6fe-f31cf31d61d7-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"198e58a9-696b-4857-a6fe-f31cf31d61d7\") " pod="openstack/mysqld-exporter-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.294904 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198e58a9-696b-4857-a6fe-f31cf31d61d7-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"198e58a9-696b-4857-a6fe-f31cf31d61d7\") " pod="openstack/mysqld-exporter-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.294995 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198e58a9-696b-4857-a6fe-f31cf31d61d7-config-data\") pod \"mysqld-exporter-0\" (UID: \"198e58a9-696b-4857-a6fe-f31cf31d61d7\") " pod="openstack/mysqld-exporter-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.308287 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.332868 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.333186 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="50b1599c-e9bb-478e-aeb9-c311d903b466" containerName="nova-api-log" containerID="cri-o://5459d74f5ea9fdeb46135081a7591d7b01e666182dfaf2c7de2d9af484a90e7e" gracePeriod=30 Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.333781 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="50b1599c-e9bb-478e-aeb9-c311d903b466" containerName="nova-api-api" containerID="cri-o://2b1e38ccb1a835a16df3418aaa3ddba896e5515e4bcd131135436b5ad05af363" gracePeriod=30 Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.396947 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/04203c31-30dc-48d1-bdb9-1a8eaf8da709-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"04203c31-30dc-48d1-bdb9-1a8eaf8da709\") " pod="openstack/kube-state-metrics-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.397014 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbkpr\" (UniqueName: \"kubernetes.io/projected/198e58a9-696b-4857-a6fe-f31cf31d61d7-kube-api-access-cbkpr\") pod \"mysqld-exporter-0\" (UID: \"198e58a9-696b-4857-a6fe-f31cf31d61d7\") " pod="openstack/mysqld-exporter-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.397082 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/198e58a9-696b-4857-a6fe-f31cf31d61d7-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"198e58a9-696b-4857-a6fe-f31cf31d61d7\") " pod="openstack/mysqld-exporter-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.397120 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/04203c31-30dc-48d1-bdb9-1a8eaf8da709-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"04203c31-30dc-48d1-bdb9-1a8eaf8da709\") " pod="openstack/kube-state-metrics-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.397173 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198e58a9-696b-4857-a6fe-f31cf31d61d7-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"198e58a9-696b-4857-a6fe-f31cf31d61d7\") " pod="openstack/mysqld-exporter-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.397225 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw5q8\" (UniqueName: \"kubernetes.io/projected/04203c31-30dc-48d1-bdb9-1a8eaf8da709-kube-api-access-xw5q8\") pod \"kube-state-metrics-0\" (UID: \"04203c31-30dc-48d1-bdb9-1a8eaf8da709\") " pod="openstack/kube-state-metrics-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.397247 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04203c31-30dc-48d1-bdb9-1a8eaf8da709-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"04203c31-30dc-48d1-bdb9-1a8eaf8da709\") " pod="openstack/kube-state-metrics-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.397300 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198e58a9-696b-4857-a6fe-f31cf31d61d7-config-data\") pod \"mysqld-exporter-0\" (UID: \"198e58a9-696b-4857-a6fe-f31cf31d61d7\") " pod="openstack/mysqld-exporter-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.407752 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/198e58a9-696b-4857-a6fe-f31cf31d61d7-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"198e58a9-696b-4857-a6fe-f31cf31d61d7\") " pod="openstack/mysqld-exporter-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.408142 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/198e58a9-696b-4857-a6fe-f31cf31d61d7-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"198e58a9-696b-4857-a6fe-f31cf31d61d7\") " pod="openstack/mysqld-exporter-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.408457 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/198e58a9-696b-4857-a6fe-f31cf31d61d7-config-data\") pod \"mysqld-exporter-0\" (UID: \"198e58a9-696b-4857-a6fe-f31cf31d61d7\") " pod="openstack/mysqld-exporter-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.444447 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbkpr\" (UniqueName: \"kubernetes.io/projected/198e58a9-696b-4857-a6fe-f31cf31d61d7-kube-api-access-cbkpr\") pod \"mysqld-exporter-0\" (UID: \"198e58a9-696b-4857-a6fe-f31cf31d61d7\") " pod="openstack/mysqld-exporter-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.499451 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/04203c31-30dc-48d1-bdb9-1a8eaf8da709-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"04203c31-30dc-48d1-bdb9-1a8eaf8da709\") " pod="openstack/kube-state-metrics-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.499620 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw5q8\" (UniqueName: \"kubernetes.io/projected/04203c31-30dc-48d1-bdb9-1a8eaf8da709-kube-api-access-xw5q8\") pod \"kube-state-metrics-0\" (UID: \"04203c31-30dc-48d1-bdb9-1a8eaf8da709\") " pod="openstack/kube-state-metrics-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.499682 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04203c31-30dc-48d1-bdb9-1a8eaf8da709-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"04203c31-30dc-48d1-bdb9-1a8eaf8da709\") " pod="openstack/kube-state-metrics-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.499827 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/04203c31-30dc-48d1-bdb9-1a8eaf8da709-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"04203c31-30dc-48d1-bdb9-1a8eaf8da709\") " pod="openstack/kube-state-metrics-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.502702 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/04203c31-30dc-48d1-bdb9-1a8eaf8da709-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"04203c31-30dc-48d1-bdb9-1a8eaf8da709\") " pod="openstack/kube-state-metrics-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.503409 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04203c31-30dc-48d1-bdb9-1a8eaf8da709-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"04203c31-30dc-48d1-bdb9-1a8eaf8da709\") " pod="openstack/kube-state-metrics-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.504449 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/04203c31-30dc-48d1-bdb9-1a8eaf8da709-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"04203c31-30dc-48d1-bdb9-1a8eaf8da709\") " pod="openstack/kube-state-metrics-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.521551 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw5q8\" (UniqueName: \"kubernetes.io/projected/04203c31-30dc-48d1-bdb9-1a8eaf8da709-kube-api-access-xw5q8\") pod \"kube-state-metrics-0\" (UID: \"04203c31-30dc-48d1-bdb9-1a8eaf8da709\") " pod="openstack/kube-state-metrics-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.579590 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.622884 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.816162 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cd541f6-4cd3-42b6-9d59-ad4b6246044e" path="/var/lib/kubelet/pods/7cd541f6-4cd3-42b6-9d59-ad4b6246044e/volumes" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.817493 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fd3d999-4b0d-422c-a7ae-01065d577c38" path="/var/lib/kubelet/pods/9fd3d999-4b0d-422c-a7ae-01065d577c38/volumes" Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.980933 4828 generic.go:334] "Generic (PLEG): container finished" podID="50b1599c-e9bb-478e-aeb9-c311d903b466" containerID="5459d74f5ea9fdeb46135081a7591d7b01e666182dfaf2c7de2d9af484a90e7e" exitCode=143 Dec 10 19:23:19 crc kubenswrapper[4828]: I1210 19:23:19.981078 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"50b1599c-e9bb-478e-aeb9-c311d903b466","Type":"ContainerDied","Data":"5459d74f5ea9fdeb46135081a7591d7b01e666182dfaf2c7de2d9af484a90e7e"} Dec 10 19:23:20 crc kubenswrapper[4828]: I1210 19:23:20.082702 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 10 19:23:20 crc kubenswrapper[4828]: I1210 19:23:20.233993 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 10 19:23:20 crc kubenswrapper[4828]: I1210 19:23:20.881746 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:23:20 crc kubenswrapper[4828]: I1210 19:23:20.949628 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-run-httpd\") pod \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " Dec 10 19:23:20 crc kubenswrapper[4828]: I1210 19:23:20.949769 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-combined-ca-bundle\") pod \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " Dec 10 19:23:20 crc kubenswrapper[4828]: I1210 19:23:20.950520 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a9cbf2a7-b956-4bf8-90f9-1987aafe08df" (UID: "a9cbf2a7-b956-4bf8-90f9-1987aafe08df"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:23:20 crc kubenswrapper[4828]: I1210 19:23:20.950597 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-scripts\") pod \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " Dec 10 19:23:20 crc kubenswrapper[4828]: I1210 19:23:20.950625 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-log-httpd\") pod \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " Dec 10 19:23:20 crc kubenswrapper[4828]: I1210 19:23:20.950665 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-sg-core-conf-yaml\") pod \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " Dec 10 19:23:20 crc kubenswrapper[4828]: I1210 19:23:20.950725 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-config-data\") pod \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " Dec 10 19:23:20 crc kubenswrapper[4828]: I1210 19:23:20.950762 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g96p\" (UniqueName: \"kubernetes.io/projected/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-kube-api-access-8g96p\") pod \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\" (UID: \"a9cbf2a7-b956-4bf8-90f9-1987aafe08df\") " Dec 10 19:23:20 crc kubenswrapper[4828]: I1210 19:23:20.951696 4828 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:20 crc kubenswrapper[4828]: I1210 19:23:20.952428 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a9cbf2a7-b956-4bf8-90f9-1987aafe08df" (UID: "a9cbf2a7-b956-4bf8-90f9-1987aafe08df"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:23:20 crc kubenswrapper[4828]: I1210 19:23:20.961994 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-scripts" (OuterVolumeSpecName: "scripts") pod "a9cbf2a7-b956-4bf8-90f9-1987aafe08df" (UID: "a9cbf2a7-b956-4bf8-90f9-1987aafe08df"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:23:20 crc kubenswrapper[4828]: I1210 19:23:20.966235 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-kube-api-access-8g96p" (OuterVolumeSpecName: "kube-api-access-8g96p") pod "a9cbf2a7-b956-4bf8-90f9-1987aafe08df" (UID: "a9cbf2a7-b956-4bf8-90f9-1987aafe08df"). InnerVolumeSpecName "kube-api-access-8g96p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.052194 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"198e58a9-696b-4857-a6fe-f31cf31d61d7","Type":"ContainerStarted","Data":"ed90c6d25f82064903958d9da51757e37c439baacf821f36a77862347d0ad530"} Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.053737 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.053759 4828 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.053771 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g96p\" (UniqueName: \"kubernetes.io/projected/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-kube-api-access-8g96p\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.080778 4828 generic.go:334] "Generic (PLEG): container finished" podID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerID="332a9ba66bc89a6c7a57b8052cb0fe701f866bb29da3b489d148e7c37821e1bb" exitCode=0 Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.080872 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9cbf2a7-b956-4bf8-90f9-1987aafe08df","Type":"ContainerDied","Data":"332a9ba66bc89a6c7a57b8052cb0fe701f866bb29da3b489d148e7c37821e1bb"} Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.080903 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a9cbf2a7-b956-4bf8-90f9-1987aafe08df","Type":"ContainerDied","Data":"7a0dcf1298239fa559381d73eb438f4eb56c57eadd348ee76ce217b269885847"} Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.080924 4828 scope.go:117] "RemoveContainer" containerID="ed4bde243a6be79ae2b68d87435d741c6a9953ed37a775ae40496dbf028d88ac" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.081097 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.090006 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a9cbf2a7-b956-4bf8-90f9-1987aafe08df" (UID: "a9cbf2a7-b956-4bf8-90f9-1987aafe08df"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.100011 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"04203c31-30dc-48d1-bdb9-1a8eaf8da709","Type":"ContainerStarted","Data":"4c823d125bef2894b48048ba7a7c97758d8218678870290cc1c499068df4fd93"} Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.101024 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.138882 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.707623871 podStartE2EDuration="2.138863183s" podCreationTimestamp="2025-12-10 19:23:19 +0000 UTC" firstStartedPulling="2025-12-10 19:23:20.24277608 +0000 UTC m=+1680.753387085" lastFinishedPulling="2025-12-10 19:23:20.674015392 +0000 UTC m=+1681.184626397" observedRunningTime="2025-12-10 19:23:21.12137765 +0000 UTC m=+1681.631988655" watchObservedRunningTime="2025-12-10 19:23:21.138863183 +0000 UTC m=+1681.649474188" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.156446 4828 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.166727 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9cbf2a7-b956-4bf8-90f9-1987aafe08df" (UID: "a9cbf2a7-b956-4bf8-90f9-1987aafe08df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.244987 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-config-data" (OuterVolumeSpecName: "config-data") pod "a9cbf2a7-b956-4bf8-90f9-1987aafe08df" (UID: "a9cbf2a7-b956-4bf8-90f9-1987aafe08df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.259125 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.259166 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9cbf2a7-b956-4bf8-90f9-1987aafe08df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.340249 4828 scope.go:117] "RemoveContainer" containerID="757b304b97d21f6ae4063f6a0334b634076c035c88a67045f183b3997fa16c51" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.372290 4828 scope.go:117] "RemoveContainer" containerID="332a9ba66bc89a6c7a57b8052cb0fe701f866bb29da3b489d148e7c37821e1bb" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.418344 4828 scope.go:117] "RemoveContainer" containerID="46bb815020c90cd623232c5f1e70bc56377d205a168df279124bb5cacf240003" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.421913 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.437225 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.447649 4828 scope.go:117] "RemoveContainer" containerID="ed4bde243a6be79ae2b68d87435d741c6a9953ed37a775ae40496dbf028d88ac" Dec 10 19:23:21 crc kubenswrapper[4828]: E1210 19:23:21.448081 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed4bde243a6be79ae2b68d87435d741c6a9953ed37a775ae40496dbf028d88ac\": container with ID starting with ed4bde243a6be79ae2b68d87435d741c6a9953ed37a775ae40496dbf028d88ac not found: ID does not exist" containerID="ed4bde243a6be79ae2b68d87435d741c6a9953ed37a775ae40496dbf028d88ac" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.448131 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed4bde243a6be79ae2b68d87435d741c6a9953ed37a775ae40496dbf028d88ac"} err="failed to get container status \"ed4bde243a6be79ae2b68d87435d741c6a9953ed37a775ae40496dbf028d88ac\": rpc error: code = NotFound desc = could not find container \"ed4bde243a6be79ae2b68d87435d741c6a9953ed37a775ae40496dbf028d88ac\": container with ID starting with ed4bde243a6be79ae2b68d87435d741c6a9953ed37a775ae40496dbf028d88ac not found: ID does not exist" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.448157 4828 scope.go:117] "RemoveContainer" containerID="757b304b97d21f6ae4063f6a0334b634076c035c88a67045f183b3997fa16c51" Dec 10 19:23:21 crc kubenswrapper[4828]: E1210 19:23:21.448468 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"757b304b97d21f6ae4063f6a0334b634076c035c88a67045f183b3997fa16c51\": container with ID starting with 757b304b97d21f6ae4063f6a0334b634076c035c88a67045f183b3997fa16c51 not found: ID does not exist" containerID="757b304b97d21f6ae4063f6a0334b634076c035c88a67045f183b3997fa16c51" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.448510 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"757b304b97d21f6ae4063f6a0334b634076c035c88a67045f183b3997fa16c51"} err="failed to get container status \"757b304b97d21f6ae4063f6a0334b634076c035c88a67045f183b3997fa16c51\": rpc error: code = NotFound desc = could not find container \"757b304b97d21f6ae4063f6a0334b634076c035c88a67045f183b3997fa16c51\": container with ID starting with 757b304b97d21f6ae4063f6a0334b634076c035c88a67045f183b3997fa16c51 not found: ID does not exist" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.448539 4828 scope.go:117] "RemoveContainer" containerID="332a9ba66bc89a6c7a57b8052cb0fe701f866bb29da3b489d148e7c37821e1bb" Dec 10 19:23:21 crc kubenswrapper[4828]: E1210 19:23:21.448725 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"332a9ba66bc89a6c7a57b8052cb0fe701f866bb29da3b489d148e7c37821e1bb\": container with ID starting with 332a9ba66bc89a6c7a57b8052cb0fe701f866bb29da3b489d148e7c37821e1bb not found: ID does not exist" containerID="332a9ba66bc89a6c7a57b8052cb0fe701f866bb29da3b489d148e7c37821e1bb" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.448748 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"332a9ba66bc89a6c7a57b8052cb0fe701f866bb29da3b489d148e7c37821e1bb"} err="failed to get container status \"332a9ba66bc89a6c7a57b8052cb0fe701f866bb29da3b489d148e7c37821e1bb\": rpc error: code = NotFound desc = could not find container \"332a9ba66bc89a6c7a57b8052cb0fe701f866bb29da3b489d148e7c37821e1bb\": container with ID starting with 332a9ba66bc89a6c7a57b8052cb0fe701f866bb29da3b489d148e7c37821e1bb not found: ID does not exist" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.448762 4828 scope.go:117] "RemoveContainer" containerID="46bb815020c90cd623232c5f1e70bc56377d205a168df279124bb5cacf240003" Dec 10 19:23:21 crc kubenswrapper[4828]: E1210 19:23:21.449054 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46bb815020c90cd623232c5f1e70bc56377d205a168df279124bb5cacf240003\": container with ID starting with 46bb815020c90cd623232c5f1e70bc56377d205a168df279124bb5cacf240003 not found: ID does not exist" containerID="46bb815020c90cd623232c5f1e70bc56377d205a168df279124bb5cacf240003" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.449076 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46bb815020c90cd623232c5f1e70bc56377d205a168df279124bb5cacf240003"} err="failed to get container status \"46bb815020c90cd623232c5f1e70bc56377d205a168df279124bb5cacf240003\": rpc error: code = NotFound desc = could not find container \"46bb815020c90cd623232c5f1e70bc56377d205a168df279124bb5cacf240003\": container with ID starting with 46bb815020c90cd623232c5f1e70bc56377d205a168df279124bb5cacf240003 not found: ID does not exist" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.450649 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:23:21 crc kubenswrapper[4828]: E1210 19:23:21.451197 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerName="sg-core" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.451213 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerName="sg-core" Dec 10 19:23:21 crc kubenswrapper[4828]: E1210 19:23:21.451223 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerName="ceilometer-notification-agent" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.451229 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerName="ceilometer-notification-agent" Dec 10 19:23:21 crc kubenswrapper[4828]: E1210 19:23:21.451245 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerName="ceilometer-central-agent" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.451254 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerName="ceilometer-central-agent" Dec 10 19:23:21 crc kubenswrapper[4828]: E1210 19:23:21.451277 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerName="proxy-httpd" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.451283 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerName="proxy-httpd" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.451530 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerName="ceilometer-notification-agent" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.451557 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerName="sg-core" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.451565 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerName="proxy-httpd" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.451595 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" containerName="ceilometer-central-agent" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.453901 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.455955 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.456525 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.456741 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.462681 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.567851 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.567992 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe3798c-c29a-4992-a6af-bc30fc137482-run-httpd\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.568060 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe3798c-c29a-4992-a6af-bc30fc137482-log-httpd\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.568253 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-scripts\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.568293 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.568322 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-257wd\" (UniqueName: \"kubernetes.io/projected/dbe3798c-c29a-4992-a6af-bc30fc137482-kube-api-access-257wd\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.568384 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-config-data\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.568419 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.670236 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.670311 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe3798c-c29a-4992-a6af-bc30fc137482-run-httpd\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.670336 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe3798c-c29a-4992-a6af-bc30fc137482-log-httpd\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.670473 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-scripts\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.670500 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.670522 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-257wd\" (UniqueName: \"kubernetes.io/projected/dbe3798c-c29a-4992-a6af-bc30fc137482-kube-api-access-257wd\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.670563 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-config-data\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.670587 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.671387 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe3798c-c29a-4992-a6af-bc30fc137482-log-httpd\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.671427 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe3798c-c29a-4992-a6af-bc30fc137482-run-httpd\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.676216 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.676859 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-config-data\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.677041 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-scripts\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.677134 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.677461 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.692055 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-257wd\" (UniqueName: \"kubernetes.io/projected/dbe3798c-c29a-4992-a6af-bc30fc137482-kube-api-access-257wd\") pod \"ceilometer-0\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.781278 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.803729 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9cbf2a7-b956-4bf8-90f9-1987aafe08df" path="/var/lib/kubelet/pods/a9cbf2a7-b956-4bf8-90f9-1987aafe08df/volumes" Dec 10 19:23:21 crc kubenswrapper[4828]: I1210 19:23:21.980261 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:23:22 crc kubenswrapper[4828]: I1210 19:23:22.116018 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"04203c31-30dc-48d1-bdb9-1a8eaf8da709","Type":"ContainerStarted","Data":"d7af2c8e7b9a772d90184b769a56684b4cfaf6fe6981ac7c6c0b037a6d71f6af"} Dec 10 19:23:22 crc kubenswrapper[4828]: I1210 19:23:22.118255 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"198e58a9-696b-4857-a6fe-f31cf31d61d7","Type":"ContainerStarted","Data":"228883df0c370b0605fe91e2d60aa307f3472842b7e1ae4c733c689ad1260e82"} Dec 10 19:23:22 crc kubenswrapper[4828]: I1210 19:23:22.138723 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=2.430472646 podStartE2EDuration="3.138687842s" podCreationTimestamp="2025-12-10 19:23:19 +0000 UTC" firstStartedPulling="2025-12-10 19:23:20.095184912 +0000 UTC m=+1680.605795917" lastFinishedPulling="2025-12-10 19:23:20.803400108 +0000 UTC m=+1681.314011113" observedRunningTime="2025-12-10 19:23:22.131291717 +0000 UTC m=+1682.641902722" watchObservedRunningTime="2025-12-10 19:23:22.138687842 +0000 UTC m=+1682.649298837" Dec 10 19:23:22 crc kubenswrapper[4828]: I1210 19:23:22.266110 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:23:22 crc kubenswrapper[4828]: W1210 19:23:22.270491 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbe3798c_c29a_4992_a6af_bc30fc137482.slice/crio-5f00918a5999b8ba678582db77a1ff3abb8735f7e2100508ecd143227cb9ec53 WatchSource:0}: Error finding container 5f00918a5999b8ba678582db77a1ff3abb8735f7e2100508ecd143227cb9ec53: Status 404 returned error can't find the container with id 5f00918a5999b8ba678582db77a1ff3abb8735f7e2100508ecd143227cb9ec53 Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.137337 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe3798c-c29a-4992-a6af-bc30fc137482","Type":"ContainerStarted","Data":"24cabad24aa98037a1da70e5a7b7f846dd8a6e7da4c4870777cc9b485dcdfea7"} Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.138202 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe3798c-c29a-4992-a6af-bc30fc137482","Type":"ContainerStarted","Data":"5f00918a5999b8ba678582db77a1ff3abb8735f7e2100508ecd143227cb9ec53"} Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.139823 4828 generic.go:334] "Generic (PLEG): container finished" podID="50b1599c-e9bb-478e-aeb9-c311d903b466" containerID="2b1e38ccb1a835a16df3418aaa3ddba896e5515e4bcd131135436b5ad05af363" exitCode=0 Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.139857 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"50b1599c-e9bb-478e-aeb9-c311d903b466","Type":"ContainerDied","Data":"2b1e38ccb1a835a16df3418aaa3ddba896e5515e4bcd131135436b5ad05af363"} Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.312994 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.411543 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50b1599c-e9bb-478e-aeb9-c311d903b466-config-data\") pod \"50b1599c-e9bb-478e-aeb9-c311d903b466\" (UID: \"50b1599c-e9bb-478e-aeb9-c311d903b466\") " Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.411646 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdjgv\" (UniqueName: \"kubernetes.io/projected/50b1599c-e9bb-478e-aeb9-c311d903b466-kube-api-access-xdjgv\") pod \"50b1599c-e9bb-478e-aeb9-c311d903b466\" (UID: \"50b1599c-e9bb-478e-aeb9-c311d903b466\") " Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.411887 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b1599c-e9bb-478e-aeb9-c311d903b466-combined-ca-bundle\") pod \"50b1599c-e9bb-478e-aeb9-c311d903b466\" (UID: \"50b1599c-e9bb-478e-aeb9-c311d903b466\") " Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.411953 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50b1599c-e9bb-478e-aeb9-c311d903b466-logs\") pod \"50b1599c-e9bb-478e-aeb9-c311d903b466\" (UID: \"50b1599c-e9bb-478e-aeb9-c311d903b466\") " Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.412860 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50b1599c-e9bb-478e-aeb9-c311d903b466-logs" (OuterVolumeSpecName: "logs") pod "50b1599c-e9bb-478e-aeb9-c311d903b466" (UID: "50b1599c-e9bb-478e-aeb9-c311d903b466"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.422745 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50b1599c-e9bb-478e-aeb9-c311d903b466-kube-api-access-xdjgv" (OuterVolumeSpecName: "kube-api-access-xdjgv") pod "50b1599c-e9bb-478e-aeb9-c311d903b466" (UID: "50b1599c-e9bb-478e-aeb9-c311d903b466"). InnerVolumeSpecName "kube-api-access-xdjgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.459865 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50b1599c-e9bb-478e-aeb9-c311d903b466-config-data" (OuterVolumeSpecName: "config-data") pod "50b1599c-e9bb-478e-aeb9-c311d903b466" (UID: "50b1599c-e9bb-478e-aeb9-c311d903b466"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.468255 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50b1599c-e9bb-478e-aeb9-c311d903b466-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "50b1599c-e9bb-478e-aeb9-c311d903b466" (UID: "50b1599c-e9bb-478e-aeb9-c311d903b466"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.514744 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b1599c-e9bb-478e-aeb9-c311d903b466-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.514791 4828 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50b1599c-e9bb-478e-aeb9-c311d903b466-logs\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.514825 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/50b1599c-e9bb-478e-aeb9-c311d903b466-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.514837 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdjgv\" (UniqueName: \"kubernetes.io/projected/50b1599c-e9bb-478e-aeb9-c311d903b466-kube-api-access-xdjgv\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:23 crc kubenswrapper[4828]: E1210 19:23:23.707835 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = parsing image configuration: Download config.json digest sha256:e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 does not match expected sha256:bed2f08e28527d1e0868dfed061398e3eeb3ad207ba3b561658b43a6ebe4d0ec" image="quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified" Dec 10 19:23:23 crc kubenswrapper[4828]: E1210 19:23:23.707981 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-notification-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68ch56h587h59ch55chd4h687h594h65hbhd8h5f9h5f9h678h597h9ch669h5b9h67dh557h575h565h675h9dh7fh59h67dh58ch564h79h66chfdq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-notification-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-257wd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/notificationhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(dbe3798c-c29a-4992-a6af-bc30fc137482): ErrImagePull: parsing image configuration: Download config.json digest sha256:e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 does not match expected sha256:bed2f08e28527d1e0868dfed061398e3eeb3ad207ba3b561658b43a6ebe4d0ec" logger="UnhandledError" Dec 10 19:23:23 crc kubenswrapper[4828]: I1210 19:23:23.789358 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:23:23 crc kubenswrapper[4828]: E1210 19:23:23.804842 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.161871 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"50b1599c-e9bb-478e-aeb9-c311d903b466","Type":"ContainerDied","Data":"1ac9631451db9ba82709ab5f4038807cf0b52336f7ec3529f83b1edecc013dbe"} Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.162209 4828 scope.go:117] "RemoveContainer" containerID="2b1e38ccb1a835a16df3418aaa3ddba896e5515e4bcd131135436b5ad05af363" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.162068 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.196046 4828 scope.go:117] "RemoveContainer" containerID="5459d74f5ea9fdeb46135081a7591d7b01e666182dfaf2c7de2d9af484a90e7e" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.201502 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.217678 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.232002 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 10 19:23:24 crc kubenswrapper[4828]: E1210 19:23:24.232748 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b1599c-e9bb-478e-aeb9-c311d903b466" containerName="nova-api-api" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.232771 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b1599c-e9bb-478e-aeb9-c311d903b466" containerName="nova-api-api" Dec 10 19:23:24 crc kubenswrapper[4828]: E1210 19:23:24.232792 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50b1599c-e9bb-478e-aeb9-c311d903b466" containerName="nova-api-log" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.232817 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="50b1599c-e9bb-478e-aeb9-c311d903b466" containerName="nova-api-log" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.233117 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b1599c-e9bb-478e-aeb9-c311d903b466" containerName="nova-api-api" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.233140 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="50b1599c-e9bb-478e-aeb9-c311d903b466" containerName="nova-api-log" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.234848 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.237708 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.239610 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.239752 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.243938 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.334839 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e880ecd-5ad3-4da8-b817-97149eff4ec2-config-data\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.334948 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e880ecd-5ad3-4da8-b817-97149eff4ec2-public-tls-certs\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.335084 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e880ecd-5ad3-4da8-b817-97149eff4ec2-logs\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.335144 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e880ecd-5ad3-4da8-b817-97149eff4ec2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.335441 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e880ecd-5ad3-4da8-b817-97149eff4ec2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.335584 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbkhs\" (UniqueName: \"kubernetes.io/projected/6e880ecd-5ad3-4da8-b817-97149eff4ec2-kube-api-access-lbkhs\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.437767 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e880ecd-5ad3-4da8-b817-97149eff4ec2-config-data\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.437894 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e880ecd-5ad3-4da8-b817-97149eff4ec2-public-tls-certs\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.437972 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e880ecd-5ad3-4da8-b817-97149eff4ec2-logs\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.438006 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e880ecd-5ad3-4da8-b817-97149eff4ec2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.438121 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e880ecd-5ad3-4da8-b817-97149eff4ec2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.438160 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbkhs\" (UniqueName: \"kubernetes.io/projected/6e880ecd-5ad3-4da8-b817-97149eff4ec2-kube-api-access-lbkhs\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.439332 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e880ecd-5ad3-4da8-b817-97149eff4ec2-logs\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.444117 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e880ecd-5ad3-4da8-b817-97149eff4ec2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.444399 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e880ecd-5ad3-4da8-b817-97149eff4ec2-config-data\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.444663 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e880ecd-5ad3-4da8-b817-97149eff4ec2-public-tls-certs\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.450170 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e880ecd-5ad3-4da8-b817-97149eff4ec2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.458405 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbkhs\" (UniqueName: \"kubernetes.io/projected/6e880ecd-5ad3-4da8-b817-97149eff4ec2-kube-api-access-lbkhs\") pod \"nova-api-0\" (UID: \"6e880ecd-5ad3-4da8-b817-97149eff4ec2\") " pod="openstack/nova-api-0" Dec 10 19:23:24 crc kubenswrapper[4828]: I1210 19:23:24.560443 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 10 19:23:25 crc kubenswrapper[4828]: I1210 19:23:25.016923 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 10 19:23:25 crc kubenswrapper[4828]: W1210 19:23:25.017917 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e880ecd_5ad3_4da8_b817_97149eff4ec2.slice/crio-3b570b3028032563a73402697002d8a41eb437beb9c65378e4480fd93a338d57 WatchSource:0}: Error finding container 3b570b3028032563a73402697002d8a41eb437beb9c65378e4480fd93a338d57: Status 404 returned error can't find the container with id 3b570b3028032563a73402697002d8a41eb437beb9c65378e4480fd93a338d57 Dec 10 19:23:25 crc kubenswrapper[4828]: I1210 19:23:25.175319 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e880ecd-5ad3-4da8-b817-97149eff4ec2","Type":"ContainerStarted","Data":"3b570b3028032563a73402697002d8a41eb437beb9c65378e4480fd93a338d57"} Dec 10 19:23:25 crc kubenswrapper[4828]: I1210 19:23:25.386001 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:23:25 crc kubenswrapper[4828]: I1210 19:23:25.483440 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-ncsqd"] Dec 10 19:23:25 crc kubenswrapper[4828]: I1210 19:23:25.483747 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" podUID="78cfbb25-b948-41f2-9157-97b4f9a70ecf" containerName="dnsmasq-dns" containerID="cri-o://9b0f0f82dec4572c981ad7cccc4373ae5c762b74bc983a788c415acf024ac396" gracePeriod=10 Dec 10 19:23:25 crc kubenswrapper[4828]: I1210 19:23:25.802449 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50b1599c-e9bb-478e-aeb9-c311d903b466" path="/var/lib/kubelet/pods/50b1599c-e9bb-478e-aeb9-c311d903b466/volumes" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.065715 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.179834 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-dns-swift-storage-0\") pod \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.179936 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqs72\" (UniqueName: \"kubernetes.io/projected/78cfbb25-b948-41f2-9157-97b4f9a70ecf-kube-api-access-rqs72\") pod \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.180021 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-ovsdbserver-sb\") pod \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.180129 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-config\") pod \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.180176 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-dns-svc\") pod \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.180204 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-ovsdbserver-nb\") pod \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\" (UID: \"78cfbb25-b948-41f2-9157-97b4f9a70ecf\") " Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.191173 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe3798c-c29a-4992-a6af-bc30fc137482","Type":"ContainerStarted","Data":"800d91f192a5b0839959fb45ccc17c0a670727b0e6dea7741381e379f5236ad2"} Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.195971 4828 generic.go:334] "Generic (PLEG): container finished" podID="78cfbb25-b948-41f2-9157-97b4f9a70ecf" containerID="9b0f0f82dec4572c981ad7cccc4373ae5c762b74bc983a788c415acf024ac396" exitCode=0 Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.196214 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" event={"ID":"78cfbb25-b948-41f2-9157-97b4f9a70ecf","Type":"ContainerDied","Data":"9b0f0f82dec4572c981ad7cccc4373ae5c762b74bc983a788c415acf024ac396"} Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.196244 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" event={"ID":"78cfbb25-b948-41f2-9157-97b4f9a70ecf","Type":"ContainerDied","Data":"4149e3cb3f6009f99de8a8433a94de03f933aa4bfc74c440b6b1b7aeda676461"} Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.196264 4828 scope.go:117] "RemoveContainer" containerID="9b0f0f82dec4572c981ad7cccc4373ae5c762b74bc983a788c415acf024ac396" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.196412 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-ncsqd" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.200120 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78cfbb25-b948-41f2-9157-97b4f9a70ecf-kube-api-access-rqs72" (OuterVolumeSpecName: "kube-api-access-rqs72") pod "78cfbb25-b948-41f2-9157-97b4f9a70ecf" (UID: "78cfbb25-b948-41f2-9157-97b4f9a70ecf"). InnerVolumeSpecName "kube-api-access-rqs72". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.202063 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e880ecd-5ad3-4da8-b817-97149eff4ec2","Type":"ContainerStarted","Data":"0f4a1bd6ba336f18bb0c196879e7047550df7c3fbad727c77771bfe2de0f8c60"} Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.202118 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e880ecd-5ad3-4da8-b817-97149eff4ec2","Type":"ContainerStarted","Data":"142b504e064446a72e543c0c1acf56c3984d51c211c38d816c7d8da1c404df04"} Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.227726 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.227706666 podStartE2EDuration="2.227706666s" podCreationTimestamp="2025-12-10 19:23:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:23:26.222219431 +0000 UTC m=+1686.732830436" watchObservedRunningTime="2025-12-10 19:23:26.227706666 +0000 UTC m=+1686.738317681" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.255483 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "78cfbb25-b948-41f2-9157-97b4f9a70ecf" (UID: "78cfbb25-b948-41f2-9157-97b4f9a70ecf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.259161 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "78cfbb25-b948-41f2-9157-97b4f9a70ecf" (UID: "78cfbb25-b948-41f2-9157-97b4f9a70ecf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.270022 4828 scope.go:117] "RemoveContainer" containerID="98d08462a5b2fd62e809965335547b0ba03a8c839194ca947320c732a43e7b54" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.270029 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-config" (OuterVolumeSpecName: "config") pod "78cfbb25-b948-41f2-9157-97b4f9a70ecf" (UID: "78cfbb25-b948-41f2-9157-97b4f9a70ecf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.271130 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "78cfbb25-b948-41f2-9157-97b4f9a70ecf" (UID: "78cfbb25-b948-41f2-9157-97b4f9a70ecf"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.274902 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "78cfbb25-b948-41f2-9157-97b4f9a70ecf" (UID: "78cfbb25-b948-41f2-9157-97b4f9a70ecf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.283286 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.284014 4828 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.284108 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.284183 4828 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.284254 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqs72\" (UniqueName: \"kubernetes.io/projected/78cfbb25-b948-41f2-9157-97b4f9a70ecf-kube-api-access-rqs72\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.284351 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78cfbb25-b948-41f2-9157-97b4f9a70ecf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.291195 4828 scope.go:117] "RemoveContainer" containerID="9b0f0f82dec4572c981ad7cccc4373ae5c762b74bc983a788c415acf024ac396" Dec 10 19:23:26 crc kubenswrapper[4828]: E1210 19:23:26.291783 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b0f0f82dec4572c981ad7cccc4373ae5c762b74bc983a788c415acf024ac396\": container with ID starting with 9b0f0f82dec4572c981ad7cccc4373ae5c762b74bc983a788c415acf024ac396 not found: ID does not exist" containerID="9b0f0f82dec4572c981ad7cccc4373ae5c762b74bc983a788c415acf024ac396" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.291832 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b0f0f82dec4572c981ad7cccc4373ae5c762b74bc983a788c415acf024ac396"} err="failed to get container status \"9b0f0f82dec4572c981ad7cccc4373ae5c762b74bc983a788c415acf024ac396\": rpc error: code = NotFound desc = could not find container \"9b0f0f82dec4572c981ad7cccc4373ae5c762b74bc983a788c415acf024ac396\": container with ID starting with 9b0f0f82dec4572c981ad7cccc4373ae5c762b74bc983a788c415acf024ac396 not found: ID does not exist" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.291860 4828 scope.go:117] "RemoveContainer" containerID="98d08462a5b2fd62e809965335547b0ba03a8c839194ca947320c732a43e7b54" Dec 10 19:23:26 crc kubenswrapper[4828]: E1210 19:23:26.292144 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98d08462a5b2fd62e809965335547b0ba03a8c839194ca947320c732a43e7b54\": container with ID starting with 98d08462a5b2fd62e809965335547b0ba03a8c839194ca947320c732a43e7b54 not found: ID does not exist" containerID="98d08462a5b2fd62e809965335547b0ba03a8c839194ca947320c732a43e7b54" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.292173 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98d08462a5b2fd62e809965335547b0ba03a8c839194ca947320c732a43e7b54"} err="failed to get container status \"98d08462a5b2fd62e809965335547b0ba03a8c839194ca947320c732a43e7b54\": rpc error: code = NotFound desc = could not find container \"98d08462a5b2fd62e809965335547b0ba03a8c839194ca947320c732a43e7b54\": container with ID starting with 98d08462a5b2fd62e809965335547b0ba03a8c839194ca947320c732a43e7b54 not found: ID does not exist" Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.544857 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-ncsqd"] Dec 10 19:23:26 crc kubenswrapper[4828]: I1210 19:23:26.553257 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-ncsqd"] Dec 10 19:23:27 crc kubenswrapper[4828]: E1210 19:23:27.020241 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-notification-agent\" with ErrImagePull: \"parsing image configuration: Download config.json digest sha256:e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855 does not match expected sha256:bed2f08e28527d1e0868dfed061398e3eeb3ad207ba3b561658b43a6ebe4d0ec\"" pod="openstack/ceilometer-0" podUID="dbe3798c-c29a-4992-a6af-bc30fc137482" Dec 10 19:23:27 crc kubenswrapper[4828]: I1210 19:23:27.215630 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe3798c-c29a-4992-a6af-bc30fc137482","Type":"ContainerStarted","Data":"e08437b20383b2212702a35c8d82bd43e53005a62c66c77523e0b938c171ba33"} Dec 10 19:23:27 crc kubenswrapper[4828]: I1210 19:23:27.215806 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbe3798c-c29a-4992-a6af-bc30fc137482" containerName="proxy-httpd" containerID="cri-o://e08437b20383b2212702a35c8d82bd43e53005a62c66c77523e0b938c171ba33" gracePeriod=30 Dec 10 19:23:27 crc kubenswrapper[4828]: I1210 19:23:27.215806 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbe3798c-c29a-4992-a6af-bc30fc137482" containerName="sg-core" containerID="cri-o://800d91f192a5b0839959fb45ccc17c0a670727b0e6dea7741381e379f5236ad2" gracePeriod=30 Dec 10 19:23:27 crc kubenswrapper[4828]: I1210 19:23:27.215852 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 19:23:27 crc kubenswrapper[4828]: I1210 19:23:27.215932 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbe3798c-c29a-4992-a6af-bc30fc137482" containerName="ceilometer-central-agent" containerID="cri-o://24cabad24aa98037a1da70e5a7b7f846dd8a6e7da4c4870777cc9b485dcdfea7" gracePeriod=30 Dec 10 19:23:27 crc kubenswrapper[4828]: I1210 19:23:27.801893 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78cfbb25-b948-41f2-9157-97b4f9a70ecf" path="/var/lib/kubelet/pods/78cfbb25-b948-41f2-9157-97b4f9a70ecf/volumes" Dec 10 19:23:28 crc kubenswrapper[4828]: I1210 19:23:28.229227 4828 generic.go:334] "Generic (PLEG): container finished" podID="dbe3798c-c29a-4992-a6af-bc30fc137482" containerID="e08437b20383b2212702a35c8d82bd43e53005a62c66c77523e0b938c171ba33" exitCode=0 Dec 10 19:23:28 crc kubenswrapper[4828]: I1210 19:23:28.229265 4828 generic.go:334] "Generic (PLEG): container finished" podID="dbe3798c-c29a-4992-a6af-bc30fc137482" containerID="800d91f192a5b0839959fb45ccc17c0a670727b0e6dea7741381e379f5236ad2" exitCode=2 Dec 10 19:23:28 crc kubenswrapper[4828]: I1210 19:23:28.229287 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe3798c-c29a-4992-a6af-bc30fc137482","Type":"ContainerDied","Data":"e08437b20383b2212702a35c8d82bd43e53005a62c66c77523e0b938c171ba33"} Dec 10 19:23:28 crc kubenswrapper[4828]: I1210 19:23:28.229318 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe3798c-c29a-4992-a6af-bc30fc137482","Type":"ContainerDied","Data":"800d91f192a5b0839959fb45ccc17c0a670727b0e6dea7741381e379f5236ad2"} Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.049980 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.148089 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe3798c-c29a-4992-a6af-bc30fc137482-run-httpd\") pod \"dbe3798c-c29a-4992-a6af-bc30fc137482\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.148165 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-combined-ca-bundle\") pod \"dbe3798c-c29a-4992-a6af-bc30fc137482\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.148248 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-ceilometer-tls-certs\") pod \"dbe3798c-c29a-4992-a6af-bc30fc137482\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.148290 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-config-data\") pod \"dbe3798c-c29a-4992-a6af-bc30fc137482\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.148335 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-257wd\" (UniqueName: \"kubernetes.io/projected/dbe3798c-c29a-4992-a6af-bc30fc137482-kube-api-access-257wd\") pod \"dbe3798c-c29a-4992-a6af-bc30fc137482\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.148369 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-sg-core-conf-yaml\") pod \"dbe3798c-c29a-4992-a6af-bc30fc137482\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.148430 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe3798c-c29a-4992-a6af-bc30fc137482-log-httpd\") pod \"dbe3798c-c29a-4992-a6af-bc30fc137482\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.148541 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-scripts\") pod \"dbe3798c-c29a-4992-a6af-bc30fc137482\" (UID: \"dbe3798c-c29a-4992-a6af-bc30fc137482\") " Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.148532 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbe3798c-c29a-4992-a6af-bc30fc137482-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dbe3798c-c29a-4992-a6af-bc30fc137482" (UID: "dbe3798c-c29a-4992-a6af-bc30fc137482"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.149019 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbe3798c-c29a-4992-a6af-bc30fc137482-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dbe3798c-c29a-4992-a6af-bc30fc137482" (UID: "dbe3798c-c29a-4992-a6af-bc30fc137482"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.149068 4828 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe3798c-c29a-4992-a6af-bc30fc137482-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.154084 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbe3798c-c29a-4992-a6af-bc30fc137482-kube-api-access-257wd" (OuterVolumeSpecName: "kube-api-access-257wd") pod "dbe3798c-c29a-4992-a6af-bc30fc137482" (UID: "dbe3798c-c29a-4992-a6af-bc30fc137482"). InnerVolumeSpecName "kube-api-access-257wd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.154297 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-scripts" (OuterVolumeSpecName: "scripts") pod "dbe3798c-c29a-4992-a6af-bc30fc137482" (UID: "dbe3798c-c29a-4992-a6af-bc30fc137482"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.183279 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "dbe3798c-c29a-4992-a6af-bc30fc137482" (UID: "dbe3798c-c29a-4992-a6af-bc30fc137482"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.213317 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dbe3798c-c29a-4992-a6af-bc30fc137482" (UID: "dbe3798c-c29a-4992-a6af-bc30fc137482"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.225695 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "dbe3798c-c29a-4992-a6af-bc30fc137482" (UID: "dbe3798c-c29a-4992-a6af-bc30fc137482"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.243685 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-config-data" (OuterVolumeSpecName: "config-data") pod "dbe3798c-c29a-4992-a6af-bc30fc137482" (UID: "dbe3798c-c29a-4992-a6af-bc30fc137482"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.250998 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-257wd\" (UniqueName: \"kubernetes.io/projected/dbe3798c-c29a-4992-a6af-bc30fc137482-kube-api-access-257wd\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.251028 4828 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.251042 4828 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbe3798c-c29a-4992-a6af-bc30fc137482-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.251054 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.251067 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.251079 4828 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.251090 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbe3798c-c29a-4992-a6af-bc30fc137482-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.252880 4828 generic.go:334] "Generic (PLEG): container finished" podID="dbe3798c-c29a-4992-a6af-bc30fc137482" containerID="24cabad24aa98037a1da70e5a7b7f846dd8a6e7da4c4870777cc9b485dcdfea7" exitCode=0 Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.252925 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe3798c-c29a-4992-a6af-bc30fc137482","Type":"ContainerDied","Data":"24cabad24aa98037a1da70e5a7b7f846dd8a6e7da4c4870777cc9b485dcdfea7"} Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.252953 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbe3798c-c29a-4992-a6af-bc30fc137482","Type":"ContainerDied","Data":"5f00918a5999b8ba678582db77a1ff3abb8735f7e2100508ecd143227cb9ec53"} Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.252973 4828 scope.go:117] "RemoveContainer" containerID="e08437b20383b2212702a35c8d82bd43e53005a62c66c77523e0b938c171ba33" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.253003 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.290271 4828 scope.go:117] "RemoveContainer" containerID="800d91f192a5b0839959fb45ccc17c0a670727b0e6dea7741381e379f5236ad2" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.324137 4828 scope.go:117] "RemoveContainer" containerID="24cabad24aa98037a1da70e5a7b7f846dd8a6e7da4c4870777cc9b485dcdfea7" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.343220 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.358760 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.364531 4828 scope.go:117] "RemoveContainer" containerID="e08437b20383b2212702a35c8d82bd43e53005a62c66c77523e0b938c171ba33" Dec 10 19:23:29 crc kubenswrapper[4828]: E1210 19:23:29.365972 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e08437b20383b2212702a35c8d82bd43e53005a62c66c77523e0b938c171ba33\": container with ID starting with e08437b20383b2212702a35c8d82bd43e53005a62c66c77523e0b938c171ba33 not found: ID does not exist" containerID="e08437b20383b2212702a35c8d82bd43e53005a62c66c77523e0b938c171ba33" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.366008 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e08437b20383b2212702a35c8d82bd43e53005a62c66c77523e0b938c171ba33"} err="failed to get container status \"e08437b20383b2212702a35c8d82bd43e53005a62c66c77523e0b938c171ba33\": rpc error: code = NotFound desc = could not find container \"e08437b20383b2212702a35c8d82bd43e53005a62c66c77523e0b938c171ba33\": container with ID starting with e08437b20383b2212702a35c8d82bd43e53005a62c66c77523e0b938c171ba33 not found: ID does not exist" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.366034 4828 scope.go:117] "RemoveContainer" containerID="800d91f192a5b0839959fb45ccc17c0a670727b0e6dea7741381e379f5236ad2" Dec 10 19:23:29 crc kubenswrapper[4828]: E1210 19:23:29.366318 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"800d91f192a5b0839959fb45ccc17c0a670727b0e6dea7741381e379f5236ad2\": container with ID starting with 800d91f192a5b0839959fb45ccc17c0a670727b0e6dea7741381e379f5236ad2 not found: ID does not exist" containerID="800d91f192a5b0839959fb45ccc17c0a670727b0e6dea7741381e379f5236ad2" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.366359 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"800d91f192a5b0839959fb45ccc17c0a670727b0e6dea7741381e379f5236ad2"} err="failed to get container status \"800d91f192a5b0839959fb45ccc17c0a670727b0e6dea7741381e379f5236ad2\": rpc error: code = NotFound desc = could not find container \"800d91f192a5b0839959fb45ccc17c0a670727b0e6dea7741381e379f5236ad2\": container with ID starting with 800d91f192a5b0839959fb45ccc17c0a670727b0e6dea7741381e379f5236ad2 not found: ID does not exist" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.366385 4828 scope.go:117] "RemoveContainer" containerID="24cabad24aa98037a1da70e5a7b7f846dd8a6e7da4c4870777cc9b485dcdfea7" Dec 10 19:23:29 crc kubenswrapper[4828]: E1210 19:23:29.366893 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24cabad24aa98037a1da70e5a7b7f846dd8a6e7da4c4870777cc9b485dcdfea7\": container with ID starting with 24cabad24aa98037a1da70e5a7b7f846dd8a6e7da4c4870777cc9b485dcdfea7 not found: ID does not exist" containerID="24cabad24aa98037a1da70e5a7b7f846dd8a6e7da4c4870777cc9b485dcdfea7" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.366919 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24cabad24aa98037a1da70e5a7b7f846dd8a6e7da4c4870777cc9b485dcdfea7"} err="failed to get container status \"24cabad24aa98037a1da70e5a7b7f846dd8a6e7da4c4870777cc9b485dcdfea7\": rpc error: code = NotFound desc = could not find container \"24cabad24aa98037a1da70e5a7b7f846dd8a6e7da4c4870777cc9b485dcdfea7\": container with ID starting with 24cabad24aa98037a1da70e5a7b7f846dd8a6e7da4c4870777cc9b485dcdfea7 not found: ID does not exist" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.367144 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:23:29 crc kubenswrapper[4828]: E1210 19:23:29.367697 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbe3798c-c29a-4992-a6af-bc30fc137482" containerName="sg-core" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.367716 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbe3798c-c29a-4992-a6af-bc30fc137482" containerName="sg-core" Dec 10 19:23:29 crc kubenswrapper[4828]: E1210 19:23:29.367734 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78cfbb25-b948-41f2-9157-97b4f9a70ecf" containerName="init" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.367741 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="78cfbb25-b948-41f2-9157-97b4f9a70ecf" containerName="init" Dec 10 19:23:29 crc kubenswrapper[4828]: E1210 19:23:29.367775 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbe3798c-c29a-4992-a6af-bc30fc137482" containerName="proxy-httpd" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.367782 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbe3798c-c29a-4992-a6af-bc30fc137482" containerName="proxy-httpd" Dec 10 19:23:29 crc kubenswrapper[4828]: E1210 19:23:29.367823 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78cfbb25-b948-41f2-9157-97b4f9a70ecf" containerName="dnsmasq-dns" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.367830 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="78cfbb25-b948-41f2-9157-97b4f9a70ecf" containerName="dnsmasq-dns" Dec 10 19:23:29 crc kubenswrapper[4828]: E1210 19:23:29.367847 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbe3798c-c29a-4992-a6af-bc30fc137482" containerName="ceilometer-central-agent" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.367852 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbe3798c-c29a-4992-a6af-bc30fc137482" containerName="ceilometer-central-agent" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.368050 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbe3798c-c29a-4992-a6af-bc30fc137482" containerName="ceilometer-central-agent" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.368071 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbe3798c-c29a-4992-a6af-bc30fc137482" containerName="sg-core" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.368082 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbe3798c-c29a-4992-a6af-bc30fc137482" containerName="proxy-httpd" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.368100 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="78cfbb25-b948-41f2-9157-97b4f9a70ecf" containerName="dnsmasq-dns" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.370499 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.375741 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.375994 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.376112 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.379226 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.456005 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.456097 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.456128 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-config-data\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.456263 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-scripts\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.456310 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7149649-bf41-4e65-8e6d-e56f5ee0b601-run-httpd\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.456360 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7149649-bf41-4e65-8e6d-e56f5ee0b601-log-httpd\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.456486 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwp6k\" (UniqueName: \"kubernetes.io/projected/c7149649-bf41-4e65-8e6d-e56f5ee0b601-kube-api-access-pwp6k\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.456514 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.558746 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.558847 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-config-data\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.558980 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-scripts\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.559000 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7149649-bf41-4e65-8e6d-e56f5ee0b601-run-httpd\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.559037 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7149649-bf41-4e65-8e6d-e56f5ee0b601-log-httpd\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.559097 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwp6k\" (UniqueName: \"kubernetes.io/projected/c7149649-bf41-4e65-8e6d-e56f5ee0b601-kube-api-access-pwp6k\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.559118 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.559217 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.559823 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7149649-bf41-4e65-8e6d-e56f5ee0b601-log-httpd\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.559856 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7149649-bf41-4e65-8e6d-e56f5ee0b601-run-httpd\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.562405 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.563167 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-scripts\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.563251 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-config-data\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.564000 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.565039 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.576577 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwp6k\" (UniqueName: \"kubernetes.io/projected/c7149649-bf41-4e65-8e6d-e56f5ee0b601-kube-api-access-pwp6k\") pod \"ceilometer-0\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.634587 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.689189 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:23:29 crc kubenswrapper[4828]: I1210 19:23:29.810139 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbe3798c-c29a-4992-a6af-bc30fc137482" path="/var/lib/kubelet/pods/dbe3798c-c29a-4992-a6af-bc30fc137482/volumes" Dec 10 19:23:30 crc kubenswrapper[4828]: I1210 19:23:30.134162 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:23:30 crc kubenswrapper[4828]: W1210 19:23:30.139266 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7149649_bf41_4e65_8e6d_e56f5ee0b601.slice/crio-de187f38b7710d3cb3221fd6ce2baecb36ed5697cd3a4b185e2bfed6838f9d7c WatchSource:0}: Error finding container de187f38b7710d3cb3221fd6ce2baecb36ed5697cd3a4b185e2bfed6838f9d7c: Status 404 returned error can't find the container with id de187f38b7710d3cb3221fd6ce2baecb36ed5697cd3a4b185e2bfed6838f9d7c Dec 10 19:23:30 crc kubenswrapper[4828]: I1210 19:23:30.266640 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7149649-bf41-4e65-8e6d-e56f5ee0b601","Type":"ContainerStarted","Data":"de187f38b7710d3cb3221fd6ce2baecb36ed5697cd3a4b185e2bfed6838f9d7c"} Dec 10 19:23:31 crc kubenswrapper[4828]: I1210 19:23:31.282990 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7149649-bf41-4e65-8e6d-e56f5ee0b601","Type":"ContainerStarted","Data":"f42cb2c388fdde726b57297ea65c4bbcc0f54a9a9577d51ab13f9dde246b5b0a"} Dec 10 19:23:32 crc kubenswrapper[4828]: I1210 19:23:32.299769 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7149649-bf41-4e65-8e6d-e56f5ee0b601","Type":"ContainerStarted","Data":"21d2f448bc28620cfff112fffb6eb0fe58aaf4ca1066a5303c275aaa21c3fea6"} Dec 10 19:23:33 crc kubenswrapper[4828]: I1210 19:23:33.314018 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7149649-bf41-4e65-8e6d-e56f5ee0b601","Type":"ContainerStarted","Data":"24ce9a8f3fe56a4d273d6d920a990d2c56e5122f0c3c153234fec77ebbf771f8"} Dec 10 19:23:34 crc kubenswrapper[4828]: I1210 19:23:34.329410 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7149649-bf41-4e65-8e6d-e56f5ee0b601","Type":"ContainerStarted","Data":"25ed84078f542cdc52b70219de0997718d9401bc12ee0b1171324538c6d255db"} Dec 10 19:23:34 crc kubenswrapper[4828]: I1210 19:23:34.330056 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 19:23:34 crc kubenswrapper[4828]: I1210 19:23:34.364866 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.041563237 podStartE2EDuration="5.364845971s" podCreationTimestamp="2025-12-10 19:23:29 +0000 UTC" firstStartedPulling="2025-12-10 19:23:30.141720715 +0000 UTC m=+1690.652331720" lastFinishedPulling="2025-12-10 19:23:33.465003449 +0000 UTC m=+1693.975614454" observedRunningTime="2025-12-10 19:23:34.352475884 +0000 UTC m=+1694.863086929" watchObservedRunningTime="2025-12-10 19:23:34.364845971 +0000 UTC m=+1694.875456986" Dec 10 19:23:34 crc kubenswrapper[4828]: I1210 19:23:34.561389 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 19:23:34 crc kubenswrapper[4828]: I1210 19:23:34.561895 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 10 19:23:35 crc kubenswrapper[4828]: I1210 19:23:35.789473 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:23:35 crc kubenswrapper[4828]: E1210 19:23:35.790161 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:23:36 crc kubenswrapper[4828]: I1210 19:23:36.053193 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6e880ecd-5ad3-4da8-b817-97149eff4ec2" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.3:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 19:23:36 crc kubenswrapper[4828]: I1210 19:23:36.053559 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6e880ecd-5ad3-4da8-b817-97149eff4ec2" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.3:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 19:23:44 crc kubenswrapper[4828]: I1210 19:23:44.568002 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 19:23:44 crc kubenswrapper[4828]: I1210 19:23:44.568619 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 10 19:23:44 crc kubenswrapper[4828]: I1210 19:23:44.568989 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 19:23:44 crc kubenswrapper[4828]: I1210 19:23:44.569024 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 10 19:23:44 crc kubenswrapper[4828]: I1210 19:23:44.575426 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 19:23:44 crc kubenswrapper[4828]: I1210 19:23:44.577090 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 10 19:23:46 crc kubenswrapper[4828]: I1210 19:23:46.790017 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:23:46 crc kubenswrapper[4828]: E1210 19:23:46.790686 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:23:57 crc kubenswrapper[4828]: I1210 19:23:57.790870 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:23:57 crc kubenswrapper[4828]: E1210 19:23:57.792298 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:23:59 crc kubenswrapper[4828]: I1210 19:23:59.698871 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 10 19:24:10 crc kubenswrapper[4828]: I1210 19:24:10.788496 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:24:10 crc kubenswrapper[4828]: E1210 19:24:10.789475 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:24:11 crc kubenswrapper[4828]: I1210 19:24:11.137942 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-2n7h2"] Dec 10 19:24:11 crc kubenswrapper[4828]: I1210 19:24:11.148402 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-2n7h2"] Dec 10 19:24:11 crc kubenswrapper[4828]: I1210 19:24:11.251209 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-2n2fz"] Dec 10 19:24:11 crc kubenswrapper[4828]: I1210 19:24:11.253261 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2n2fz" Dec 10 19:24:11 crc kubenswrapper[4828]: I1210 19:24:11.264703 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-2n2fz"] Dec 10 19:24:11 crc kubenswrapper[4828]: I1210 19:24:11.321792 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2449558a-d727-461c-acb2-99e380868fb1-config-data\") pod \"heat-db-sync-2n2fz\" (UID: \"2449558a-d727-461c-acb2-99e380868fb1\") " pod="openstack/heat-db-sync-2n2fz" Dec 10 19:24:11 crc kubenswrapper[4828]: I1210 19:24:11.322147 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvsmv\" (UniqueName: \"kubernetes.io/projected/2449558a-d727-461c-acb2-99e380868fb1-kube-api-access-vvsmv\") pod \"heat-db-sync-2n2fz\" (UID: \"2449558a-d727-461c-acb2-99e380868fb1\") " pod="openstack/heat-db-sync-2n2fz" Dec 10 19:24:11 crc kubenswrapper[4828]: I1210 19:24:11.322580 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2449558a-d727-461c-acb2-99e380868fb1-combined-ca-bundle\") pod \"heat-db-sync-2n2fz\" (UID: \"2449558a-d727-461c-acb2-99e380868fb1\") " pod="openstack/heat-db-sync-2n2fz" Dec 10 19:24:11 crc kubenswrapper[4828]: I1210 19:24:11.424522 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2449558a-d727-461c-acb2-99e380868fb1-combined-ca-bundle\") pod \"heat-db-sync-2n2fz\" (UID: \"2449558a-d727-461c-acb2-99e380868fb1\") " pod="openstack/heat-db-sync-2n2fz" Dec 10 19:24:11 crc kubenswrapper[4828]: I1210 19:24:11.424645 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2449558a-d727-461c-acb2-99e380868fb1-config-data\") pod \"heat-db-sync-2n2fz\" (UID: \"2449558a-d727-461c-acb2-99e380868fb1\") " pod="openstack/heat-db-sync-2n2fz" Dec 10 19:24:11 crc kubenswrapper[4828]: I1210 19:24:11.424679 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvsmv\" (UniqueName: \"kubernetes.io/projected/2449558a-d727-461c-acb2-99e380868fb1-kube-api-access-vvsmv\") pod \"heat-db-sync-2n2fz\" (UID: \"2449558a-d727-461c-acb2-99e380868fb1\") " pod="openstack/heat-db-sync-2n2fz" Dec 10 19:24:11 crc kubenswrapper[4828]: I1210 19:24:11.431970 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2449558a-d727-461c-acb2-99e380868fb1-config-data\") pod \"heat-db-sync-2n2fz\" (UID: \"2449558a-d727-461c-acb2-99e380868fb1\") " pod="openstack/heat-db-sync-2n2fz" Dec 10 19:24:11 crc kubenswrapper[4828]: I1210 19:24:11.439856 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2449558a-d727-461c-acb2-99e380868fb1-combined-ca-bundle\") pod \"heat-db-sync-2n2fz\" (UID: \"2449558a-d727-461c-acb2-99e380868fb1\") " pod="openstack/heat-db-sync-2n2fz" Dec 10 19:24:11 crc kubenswrapper[4828]: I1210 19:24:11.442342 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvsmv\" (UniqueName: \"kubernetes.io/projected/2449558a-d727-461c-acb2-99e380868fb1-kube-api-access-vvsmv\") pod \"heat-db-sync-2n2fz\" (UID: \"2449558a-d727-461c-acb2-99e380868fb1\") " pod="openstack/heat-db-sync-2n2fz" Dec 10 19:24:11 crc kubenswrapper[4828]: I1210 19:24:11.584570 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2n2fz" Dec 10 19:24:11 crc kubenswrapper[4828]: I1210 19:24:11.808629 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acbabd11-f05a-4d18-a19a-6f1a74212a65" path="/var/lib/kubelet/pods/acbabd11-f05a-4d18-a19a-6f1a74212a65/volumes" Dec 10 19:24:12 crc kubenswrapper[4828]: I1210 19:24:12.102615 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-2n2fz"] Dec 10 19:24:12 crc kubenswrapper[4828]: I1210 19:24:12.757977 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2n2fz" event={"ID":"2449558a-d727-461c-acb2-99e380868fb1","Type":"ContainerStarted","Data":"fe3b73df95e072c9b7023f992c751176eb0e8b0d2d9f21d6a81247fddbef3cc0"} Dec 10 19:24:13 crc kubenswrapper[4828]: I1210 19:24:13.139526 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:24:13 crc kubenswrapper[4828]: I1210 19:24:13.146518 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerName="ceilometer-central-agent" containerID="cri-o://f42cb2c388fdde726b57297ea65c4bbcc0f54a9a9577d51ab13f9dde246b5b0a" gracePeriod=30 Dec 10 19:24:13 crc kubenswrapper[4828]: I1210 19:24:13.146984 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerName="proxy-httpd" containerID="cri-o://25ed84078f542cdc52b70219de0997718d9401bc12ee0b1171324538c6d255db" gracePeriod=30 Dec 10 19:24:13 crc kubenswrapper[4828]: I1210 19:24:13.147052 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerName="sg-core" containerID="cri-o://24ce9a8f3fe56a4d273d6d920a990d2c56e5122f0c3c153234fec77ebbf771f8" gracePeriod=30 Dec 10 19:24:13 crc kubenswrapper[4828]: I1210 19:24:13.147149 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerName="ceilometer-notification-agent" containerID="cri-o://21d2f448bc28620cfff112fffb6eb0fe58aaf4ca1066a5303c275aaa21c3fea6" gracePeriod=30 Dec 10 19:24:13 crc kubenswrapper[4828]: I1210 19:24:13.778903 4828 generic.go:334] "Generic (PLEG): container finished" podID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerID="25ed84078f542cdc52b70219de0997718d9401bc12ee0b1171324538c6d255db" exitCode=0 Dec 10 19:24:13 crc kubenswrapper[4828]: I1210 19:24:13.779210 4828 generic.go:334] "Generic (PLEG): container finished" podID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerID="24ce9a8f3fe56a4d273d6d920a990d2c56e5122f0c3c153234fec77ebbf771f8" exitCode=2 Dec 10 19:24:13 crc kubenswrapper[4828]: I1210 19:24:13.779218 4828 generic.go:334] "Generic (PLEG): container finished" podID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerID="f42cb2c388fdde726b57297ea65c4bbcc0f54a9a9577d51ab13f9dde246b5b0a" exitCode=0 Dec 10 19:24:13 crc kubenswrapper[4828]: I1210 19:24:13.778965 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7149649-bf41-4e65-8e6d-e56f5ee0b601","Type":"ContainerDied","Data":"25ed84078f542cdc52b70219de0997718d9401bc12ee0b1171324538c6d255db"} Dec 10 19:24:13 crc kubenswrapper[4828]: I1210 19:24:13.779254 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7149649-bf41-4e65-8e6d-e56f5ee0b601","Type":"ContainerDied","Data":"24ce9a8f3fe56a4d273d6d920a990d2c56e5122f0c3c153234fec77ebbf771f8"} Dec 10 19:24:13 crc kubenswrapper[4828]: I1210 19:24:13.779269 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7149649-bf41-4e65-8e6d-e56f5ee0b601","Type":"ContainerDied","Data":"f42cb2c388fdde726b57297ea65c4bbcc0f54a9a9577d51ab13f9dde246b5b0a"} Dec 10 19:24:13 crc kubenswrapper[4828]: I1210 19:24:13.855283 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:24:13 crc kubenswrapper[4828]: I1210 19:24:13.934828 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.496829 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.605441 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7149649-bf41-4e65-8e6d-e56f5ee0b601-run-httpd\") pod \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.605820 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-scripts\") pod \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.605889 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-sg-core-conf-yaml\") pod \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.605899 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7149649-bf41-4e65-8e6d-e56f5ee0b601-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c7149649-bf41-4e65-8e6d-e56f5ee0b601" (UID: "c7149649-bf41-4e65-8e6d-e56f5ee0b601"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.605957 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-combined-ca-bundle\") pod \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.605987 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwp6k\" (UniqueName: \"kubernetes.io/projected/c7149649-bf41-4e65-8e6d-e56f5ee0b601-kube-api-access-pwp6k\") pod \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.606022 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-config-data\") pod \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.606055 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7149649-bf41-4e65-8e6d-e56f5ee0b601-log-httpd\") pod \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.606108 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-ceilometer-tls-certs\") pod \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\" (UID: \"c7149649-bf41-4e65-8e6d-e56f5ee0b601\") " Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.606845 4828 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7149649-bf41-4e65-8e6d-e56f5ee0b601-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.607750 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7149649-bf41-4e65-8e6d-e56f5ee0b601-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c7149649-bf41-4e65-8e6d-e56f5ee0b601" (UID: "c7149649-bf41-4e65-8e6d-e56f5ee0b601"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.613280 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-scripts" (OuterVolumeSpecName: "scripts") pod "c7149649-bf41-4e65-8e6d-e56f5ee0b601" (UID: "c7149649-bf41-4e65-8e6d-e56f5ee0b601"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.647810 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c7149649-bf41-4e65-8e6d-e56f5ee0b601" (UID: "c7149649-bf41-4e65-8e6d-e56f5ee0b601"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.650863 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7149649-bf41-4e65-8e6d-e56f5ee0b601-kube-api-access-pwp6k" (OuterVolumeSpecName: "kube-api-access-pwp6k") pod "c7149649-bf41-4e65-8e6d-e56f5ee0b601" (UID: "c7149649-bf41-4e65-8e6d-e56f5ee0b601"). InnerVolumeSpecName "kube-api-access-pwp6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.709349 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwp6k\" (UniqueName: \"kubernetes.io/projected/c7149649-bf41-4e65-8e6d-e56f5ee0b601-kube-api-access-pwp6k\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.709380 4828 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7149649-bf41-4e65-8e6d-e56f5ee0b601-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.709389 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.709399 4828 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.745267 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "c7149649-bf41-4e65-8e6d-e56f5ee0b601" (UID: "c7149649-bf41-4e65-8e6d-e56f5ee0b601"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.751540 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c7149649-bf41-4e65-8e6d-e56f5ee0b601" (UID: "c7149649-bf41-4e65-8e6d-e56f5ee0b601"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.802294 4828 generic.go:334] "Generic (PLEG): container finished" podID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerID="21d2f448bc28620cfff112fffb6eb0fe58aaf4ca1066a5303c275aaa21c3fea6" exitCode=0 Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.802566 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.802965 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7149649-bf41-4e65-8e6d-e56f5ee0b601","Type":"ContainerDied","Data":"21d2f448bc28620cfff112fffb6eb0fe58aaf4ca1066a5303c275aaa21c3fea6"} Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.802992 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7149649-bf41-4e65-8e6d-e56f5ee0b601","Type":"ContainerDied","Data":"de187f38b7710d3cb3221fd6ce2baecb36ed5697cd3a4b185e2bfed6838f9d7c"} Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.803007 4828 scope.go:117] "RemoveContainer" containerID="25ed84078f542cdc52b70219de0997718d9401bc12ee0b1171324538c6d255db" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.811465 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.811502 4828 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.816773 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-config-data" (OuterVolumeSpecName: "config-data") pod "c7149649-bf41-4e65-8e6d-e56f5ee0b601" (UID: "c7149649-bf41-4e65-8e6d-e56f5ee0b601"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.841134 4828 scope.go:117] "RemoveContainer" containerID="24ce9a8f3fe56a4d273d6d920a990d2c56e5122f0c3c153234fec77ebbf771f8" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.874200 4828 scope.go:117] "RemoveContainer" containerID="21d2f448bc28620cfff112fffb6eb0fe58aaf4ca1066a5303c275aaa21c3fea6" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.909082 4828 scope.go:117] "RemoveContainer" containerID="f42cb2c388fdde726b57297ea65c4bbcc0f54a9a9577d51ab13f9dde246b5b0a" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.913389 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7149649-bf41-4e65-8e6d-e56f5ee0b601-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.973451 4828 scope.go:117] "RemoveContainer" containerID="25ed84078f542cdc52b70219de0997718d9401bc12ee0b1171324538c6d255db" Dec 10 19:24:14 crc kubenswrapper[4828]: E1210 19:24:14.976284 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25ed84078f542cdc52b70219de0997718d9401bc12ee0b1171324538c6d255db\": container with ID starting with 25ed84078f542cdc52b70219de0997718d9401bc12ee0b1171324538c6d255db not found: ID does not exist" containerID="25ed84078f542cdc52b70219de0997718d9401bc12ee0b1171324538c6d255db" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.976328 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25ed84078f542cdc52b70219de0997718d9401bc12ee0b1171324538c6d255db"} err="failed to get container status \"25ed84078f542cdc52b70219de0997718d9401bc12ee0b1171324538c6d255db\": rpc error: code = NotFound desc = could not find container \"25ed84078f542cdc52b70219de0997718d9401bc12ee0b1171324538c6d255db\": container with ID starting with 25ed84078f542cdc52b70219de0997718d9401bc12ee0b1171324538c6d255db not found: ID does not exist" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.976354 4828 scope.go:117] "RemoveContainer" containerID="24ce9a8f3fe56a4d273d6d920a990d2c56e5122f0c3c153234fec77ebbf771f8" Dec 10 19:24:14 crc kubenswrapper[4828]: E1210 19:24:14.977599 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24ce9a8f3fe56a4d273d6d920a990d2c56e5122f0c3c153234fec77ebbf771f8\": container with ID starting with 24ce9a8f3fe56a4d273d6d920a990d2c56e5122f0c3c153234fec77ebbf771f8 not found: ID does not exist" containerID="24ce9a8f3fe56a4d273d6d920a990d2c56e5122f0c3c153234fec77ebbf771f8" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.977628 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24ce9a8f3fe56a4d273d6d920a990d2c56e5122f0c3c153234fec77ebbf771f8"} err="failed to get container status \"24ce9a8f3fe56a4d273d6d920a990d2c56e5122f0c3c153234fec77ebbf771f8\": rpc error: code = NotFound desc = could not find container \"24ce9a8f3fe56a4d273d6d920a990d2c56e5122f0c3c153234fec77ebbf771f8\": container with ID starting with 24ce9a8f3fe56a4d273d6d920a990d2c56e5122f0c3c153234fec77ebbf771f8 not found: ID does not exist" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.977654 4828 scope.go:117] "RemoveContainer" containerID="21d2f448bc28620cfff112fffb6eb0fe58aaf4ca1066a5303c275aaa21c3fea6" Dec 10 19:24:14 crc kubenswrapper[4828]: E1210 19:24:14.978062 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21d2f448bc28620cfff112fffb6eb0fe58aaf4ca1066a5303c275aaa21c3fea6\": container with ID starting with 21d2f448bc28620cfff112fffb6eb0fe58aaf4ca1066a5303c275aaa21c3fea6 not found: ID does not exist" containerID="21d2f448bc28620cfff112fffb6eb0fe58aaf4ca1066a5303c275aaa21c3fea6" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.978108 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21d2f448bc28620cfff112fffb6eb0fe58aaf4ca1066a5303c275aaa21c3fea6"} err="failed to get container status \"21d2f448bc28620cfff112fffb6eb0fe58aaf4ca1066a5303c275aaa21c3fea6\": rpc error: code = NotFound desc = could not find container \"21d2f448bc28620cfff112fffb6eb0fe58aaf4ca1066a5303c275aaa21c3fea6\": container with ID starting with 21d2f448bc28620cfff112fffb6eb0fe58aaf4ca1066a5303c275aaa21c3fea6 not found: ID does not exist" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.978140 4828 scope.go:117] "RemoveContainer" containerID="f42cb2c388fdde726b57297ea65c4bbcc0f54a9a9577d51ab13f9dde246b5b0a" Dec 10 19:24:14 crc kubenswrapper[4828]: E1210 19:24:14.979422 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f42cb2c388fdde726b57297ea65c4bbcc0f54a9a9577d51ab13f9dde246b5b0a\": container with ID starting with f42cb2c388fdde726b57297ea65c4bbcc0f54a9a9577d51ab13f9dde246b5b0a not found: ID does not exist" containerID="f42cb2c388fdde726b57297ea65c4bbcc0f54a9a9577d51ab13f9dde246b5b0a" Dec 10 19:24:14 crc kubenswrapper[4828]: I1210 19:24:14.979461 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f42cb2c388fdde726b57297ea65c4bbcc0f54a9a9577d51ab13f9dde246b5b0a"} err="failed to get container status \"f42cb2c388fdde726b57297ea65c4bbcc0f54a9a9577d51ab13f9dde246b5b0a\": rpc error: code = NotFound desc = could not find container \"f42cb2c388fdde726b57297ea65c4bbcc0f54a9a9577d51ab13f9dde246b5b0a\": container with ID starting with f42cb2c388fdde726b57297ea65c4bbcc0f54a9a9577d51ab13f9dde246b5b0a not found: ID does not exist" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.152146 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.169481 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.182199 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:24:15 crc kubenswrapper[4828]: E1210 19:24:15.182821 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerName="ceilometer-central-agent" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.182844 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerName="ceilometer-central-agent" Dec 10 19:24:15 crc kubenswrapper[4828]: E1210 19:24:15.182877 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerName="sg-core" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.182886 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerName="sg-core" Dec 10 19:24:15 crc kubenswrapper[4828]: E1210 19:24:15.182913 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerName="ceilometer-notification-agent" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.182923 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerName="ceilometer-notification-agent" Dec 10 19:24:15 crc kubenswrapper[4828]: E1210 19:24:15.182934 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerName="proxy-httpd" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.182941 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerName="proxy-httpd" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.183258 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerName="sg-core" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.183299 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerName="ceilometer-central-agent" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.183314 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerName="proxy-httpd" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.183324 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" containerName="ceilometer-notification-agent" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.186638 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.204393 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.205095 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.205280 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.210695 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.321104 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-config-data\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.321220 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw4md\" (UniqueName: \"kubernetes.io/projected/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-kube-api-access-qw4md\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.321247 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-run-httpd\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.321269 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-log-httpd\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.321371 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.321641 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.321683 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.321727 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-scripts\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.424691 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.424735 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.424768 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-scripts\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.424846 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-config-data\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.424906 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw4md\" (UniqueName: \"kubernetes.io/projected/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-kube-api-access-qw4md\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.424931 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-run-httpd\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.424953 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-log-httpd\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.424993 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.426160 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-run-httpd\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.426468 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-log-httpd\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.435625 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.435624 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-scripts\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.436201 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.440391 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.447618 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-config-data\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.462494 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw4md\" (UniqueName: \"kubernetes.io/projected/cd0c6bc2-1b33-4c82-ae2f-9ea265564b82-kube-api-access-qw4md\") pod \"ceilometer-0\" (UID: \"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82\") " pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.517595 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 10 19:24:15 crc kubenswrapper[4828]: I1210 19:24:15.806351 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7149649-bf41-4e65-8e6d-e56f5ee0b601" path="/var/lib/kubelet/pods/c7149649-bf41-4e65-8e6d-e56f5ee0b601/volumes" Dec 10 19:24:16 crc kubenswrapper[4828]: I1210 19:24:16.085667 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 10 19:24:16 crc kubenswrapper[4828]: I1210 19:24:16.840580 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82","Type":"ContainerStarted","Data":"e896b4fd4544850505dbe0b27164d518c599dc514aa68d2e7f9e47f345ff33c4"} Dec 10 19:24:18 crc kubenswrapper[4828]: I1210 19:24:18.518444 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="fd6f7666-6cc8-478d-9acf-21cfbad541c9" containerName="rabbitmq" containerID="cri-o://3cf3999e422c09e05f979b2209c544abc8e36d082b76293e4209e83d892753ba" gracePeriod=604796 Dec 10 19:24:19 crc kubenswrapper[4828]: I1210 19:24:19.523160 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="b24f93de-d604-499b-be8e-2e934d119192" containerName="rabbitmq" containerID="cri-o://cebed91d9e42c67e258a49b3b4f0266cf93e32df764a470725ba35c027b4f1af" gracePeriod=604796 Dec 10 19:24:22 crc kubenswrapper[4828]: I1210 19:24:22.568949 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="fd6f7666-6cc8-478d-9acf-21cfbad541c9" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.129:5671: connect: connection refused" Dec 10 19:24:22 crc kubenswrapper[4828]: I1210 19:24:22.655466 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="b24f93de-d604-499b-be8e-2e934d119192" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.130:5671: connect: connection refused" Dec 10 19:24:23 crc kubenswrapper[4828]: I1210 19:24:23.789003 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:24:23 crc kubenswrapper[4828]: E1210 19:24:23.789616 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:24:26 crc kubenswrapper[4828]: I1210 19:24:26.980123 4828 generic.go:334] "Generic (PLEG): container finished" podID="fd6f7666-6cc8-478d-9acf-21cfbad541c9" containerID="3cf3999e422c09e05f979b2209c544abc8e36d082b76293e4209e83d892753ba" exitCode=0 Dec 10 19:24:26 crc kubenswrapper[4828]: I1210 19:24:26.980183 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fd6f7666-6cc8-478d-9acf-21cfbad541c9","Type":"ContainerDied","Data":"3cf3999e422c09e05f979b2209c544abc8e36d082b76293e4209e83d892753ba"} Dec 10 19:24:26 crc kubenswrapper[4828]: I1210 19:24:26.983368 4828 generic.go:334] "Generic (PLEG): container finished" podID="b24f93de-d604-499b-be8e-2e934d119192" containerID="cebed91d9e42c67e258a49b3b4f0266cf93e32df764a470725ba35c027b4f1af" exitCode=0 Dec 10 19:24:26 crc kubenswrapper[4828]: I1210 19:24:26.983426 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b24f93de-d604-499b-be8e-2e934d119192","Type":"ContainerDied","Data":"cebed91d9e42c67e258a49b3b4f0266cf93e32df764a470725ba35c027b4f1af"} Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.600895 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-k6pwv"] Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.607005 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.624296 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.658380 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-k6pwv"] Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.763063 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.763316 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.763442 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-config\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.763579 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.763663 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlf6b\" (UniqueName: \"kubernetes.io/projected/74792a2b-341a-4065-93e3-215c325db10c-kube-api-access-zlf6b\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.763750 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.764095 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.865806 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.865945 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.865993 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.866019 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.866065 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-config\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.866167 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.866185 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlf6b\" (UniqueName: \"kubernetes.io/projected/74792a2b-341a-4065-93e3-215c325db10c-kube-api-access-zlf6b\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.867459 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.867487 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-config\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.868030 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.868051 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.869390 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.869405 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.904621 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlf6b\" (UniqueName: \"kubernetes.io/projected/74792a2b-341a-4065-93e3-215c325db10c-kube-api-access-zlf6b\") pod \"dnsmasq-dns-5b75489c6f-k6pwv\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:28 crc kubenswrapper[4828]: I1210 19:24:28.935718 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:35 crc kubenswrapper[4828]: E1210 19:24:35.939753 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 10 19:24:35 crc kubenswrapper[4828]: E1210 19:24:35.940364 4828 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 10 19:24:35 crc kubenswrapper[4828]: E1210 19:24:35.940483 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vvsmv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-2n2fz_openstack(2449558a-d727-461c-acb2-99e380868fb1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:24:35 crc kubenswrapper[4828]: E1210 19:24:35.941653 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-2n2fz" podUID="2449558a-d727-461c-acb2-99e380868fb1" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.028933 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.039303 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.126148 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b24f93de-d604-499b-be8e-2e934d119192","Type":"ContainerDied","Data":"81896e0a70a4edf55b884889ee4445be770f759065a5aa263cbeb0bda75df76c"} Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.126206 4828 scope.go:117] "RemoveContainer" containerID="cebed91d9e42c67e258a49b3b4f0266cf93e32df764a470725ba35c027b4f1af" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.126370 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.134019 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fd6f7666-6cc8-478d-9acf-21cfbad541c9","Type":"ContainerDied","Data":"4680091c3e0313bdbe5d56a5292605397c48c1f482b86a0c2b6c8c0514d3d36d"} Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.134311 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: E1210 19:24:36.136153 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-2n2fz" podUID="2449558a-d727-461c-acb2-99e380868fb1" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.148702 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hftq\" (UniqueName: \"kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-kube-api-access-7hftq\") pod \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.148785 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.148875 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-plugins-conf\") pod \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.148910 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-plugins\") pod \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.148938 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-confd\") pod \"b24f93de-d604-499b-be8e-2e934d119192\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.148957 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fd6f7666-6cc8-478d-9acf-21cfbad541c9-erlang-cookie-secret\") pod \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.149013 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-plugins-conf\") pod \"b24f93de-d604-499b-be8e-2e934d119192\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.149039 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-erlang-cookie\") pod \"b24f93de-d604-499b-be8e-2e934d119192\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.149053 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-tls\") pod \"b24f93de-d604-499b-be8e-2e934d119192\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.149081 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"b24f93de-d604-499b-be8e-2e934d119192\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.149124 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-config-data\") pod \"b24f93de-d604-499b-be8e-2e934d119192\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.149165 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-erlang-cookie\") pod \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.149189 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6jxn8\" (UniqueName: \"kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-kube-api-access-6jxn8\") pod \"b24f93de-d604-499b-be8e-2e934d119192\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.149223 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-server-conf\") pod \"b24f93de-d604-499b-be8e-2e934d119192\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.149267 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-config-data\") pod \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.149286 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-server-conf\") pod \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.149321 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b24f93de-d604-499b-be8e-2e934d119192-erlang-cookie-secret\") pod \"b24f93de-d604-499b-be8e-2e934d119192\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.149354 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-confd\") pod \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.149371 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-plugins\") pod \"b24f93de-d604-499b-be8e-2e934d119192\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.149396 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b24f93de-d604-499b-be8e-2e934d119192-pod-info\") pod \"b24f93de-d604-499b-be8e-2e934d119192\" (UID: \"b24f93de-d604-499b-be8e-2e934d119192\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.149431 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-tls\") pod \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.149495 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fd6f7666-6cc8-478d-9acf-21cfbad541c9-pod-info\") pod \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\" (UID: \"fd6f7666-6cc8-478d-9acf-21cfbad541c9\") " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.151448 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "fd6f7666-6cc8-478d-9acf-21cfbad541c9" (UID: "fd6f7666-6cc8-478d-9acf-21cfbad541c9"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.152672 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "b24f93de-d604-499b-be8e-2e934d119192" (UID: "b24f93de-d604-499b-be8e-2e934d119192"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.158555 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "b24f93de-d604-499b-be8e-2e934d119192" (UID: "b24f93de-d604-499b-be8e-2e934d119192"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.162196 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "fd6f7666-6cc8-478d-9acf-21cfbad541c9" (UID: "fd6f7666-6cc8-478d-9acf-21cfbad541c9"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.162567 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b24f93de-d604-499b-be8e-2e934d119192" (UID: "b24f93de-d604-499b-be8e-2e934d119192"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.169224 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "fd6f7666-6cc8-478d-9acf-21cfbad541c9" (UID: "fd6f7666-6cc8-478d-9acf-21cfbad541c9"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.169314 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/fd6f7666-6cc8-478d-9acf-21cfbad541c9-pod-info" (OuterVolumeSpecName: "pod-info") pod "fd6f7666-6cc8-478d-9acf-21cfbad541c9" (UID: "fd6f7666-6cc8-478d-9acf-21cfbad541c9"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.169385 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "fd6f7666-6cc8-478d-9acf-21cfbad541c9" (UID: "fd6f7666-6cc8-478d-9acf-21cfbad541c9"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.169641 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "b24f93de-d604-499b-be8e-2e934d119192" (UID: "b24f93de-d604-499b-be8e-2e934d119192"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.170243 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "fd6f7666-6cc8-478d-9acf-21cfbad541c9" (UID: "fd6f7666-6cc8-478d-9acf-21cfbad541c9"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.174407 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "b24f93de-d604-499b-be8e-2e934d119192" (UID: "b24f93de-d604-499b-be8e-2e934d119192"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.174699 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b24f93de-d604-499b-be8e-2e934d119192-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "b24f93de-d604-499b-be8e-2e934d119192" (UID: "b24f93de-d604-499b-be8e-2e934d119192"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.175315 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd6f7666-6cc8-478d-9acf-21cfbad541c9-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "fd6f7666-6cc8-478d-9acf-21cfbad541c9" (UID: "fd6f7666-6cc8-478d-9acf-21cfbad541c9"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.177130 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-kube-api-access-6jxn8" (OuterVolumeSpecName: "kube-api-access-6jxn8") pod "b24f93de-d604-499b-be8e-2e934d119192" (UID: "b24f93de-d604-499b-be8e-2e934d119192"). InnerVolumeSpecName "kube-api-access-6jxn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.182099 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/b24f93de-d604-499b-be8e-2e934d119192-pod-info" (OuterVolumeSpecName: "pod-info") pod "b24f93de-d604-499b-be8e-2e934d119192" (UID: "b24f93de-d604-499b-be8e-2e934d119192"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.182213 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-kube-api-access-7hftq" (OuterVolumeSpecName: "kube-api-access-7hftq") pod "fd6f7666-6cc8-478d-9acf-21cfbad541c9" (UID: "fd6f7666-6cc8-478d-9acf-21cfbad541c9"). InnerVolumeSpecName "kube-api-access-7hftq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.207965 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-config-data" (OuterVolumeSpecName: "config-data") pod "fd6f7666-6cc8-478d-9acf-21cfbad541c9" (UID: "fd6f7666-6cc8-478d-9acf-21cfbad541c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.252413 4828 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fd6f7666-6cc8-478d-9acf-21cfbad541c9-pod-info\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.252449 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hftq\" (UniqueName: \"kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-kube-api-access-7hftq\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.252471 4828 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.252480 4828 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.252490 4828 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.252500 4828 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fd6f7666-6cc8-478d-9acf-21cfbad541c9-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.252580 4828 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.252589 4828 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.252598 4828 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.252612 4828 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.252620 4828 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.252628 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6jxn8\" (UniqueName: \"kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-kube-api-access-6jxn8\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.252636 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.252645 4828 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b24f93de-d604-499b-be8e-2e934d119192-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.252654 4828 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.252665 4828 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b24f93de-d604-499b-be8e-2e934d119192-pod-info\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.252674 4828 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.278387 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-config-data" (OuterVolumeSpecName: "config-data") pod "b24f93de-d604-499b-be8e-2e934d119192" (UID: "b24f93de-d604-499b-be8e-2e934d119192"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.297714 4828 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.329624 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-server-conf" (OuterVolumeSpecName: "server-conf") pod "fd6f7666-6cc8-478d-9acf-21cfbad541c9" (UID: "fd6f7666-6cc8-478d-9acf-21cfbad541c9"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.330828 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-server-conf" (OuterVolumeSpecName: "server-conf") pod "b24f93de-d604-499b-be8e-2e934d119192" (UID: "b24f93de-d604-499b-be8e-2e934d119192"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.352596 4828 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.356117 4828 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.356145 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.356158 4828 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b24f93de-d604-499b-be8e-2e934d119192-server-conf\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.356166 4828 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fd6f7666-6cc8-478d-9acf-21cfbad541c9-server-conf\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.356174 4828 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.441589 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "b24f93de-d604-499b-be8e-2e934d119192" (UID: "b24f93de-d604-499b-be8e-2e934d119192"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.458540 4828 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b24f93de-d604-499b-be8e-2e934d119192-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.461086 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "fd6f7666-6cc8-478d-9acf-21cfbad541c9" (UID: "fd6f7666-6cc8-478d-9acf-21cfbad541c9"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.565578 4828 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fd6f7666-6cc8-478d-9acf-21cfbad541c9-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.776383 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.796105 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.818996 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.851762 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.866990 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:24:36 crc kubenswrapper[4828]: E1210 19:24:36.867617 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd6f7666-6cc8-478d-9acf-21cfbad541c9" containerName="setup-container" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.867640 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd6f7666-6cc8-478d-9acf-21cfbad541c9" containerName="setup-container" Dec 10 19:24:36 crc kubenswrapper[4828]: E1210 19:24:36.867654 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b24f93de-d604-499b-be8e-2e934d119192" containerName="setup-container" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.867661 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="b24f93de-d604-499b-be8e-2e934d119192" containerName="setup-container" Dec 10 19:24:36 crc kubenswrapper[4828]: E1210 19:24:36.867678 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd6f7666-6cc8-478d-9acf-21cfbad541c9" containerName="rabbitmq" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.867685 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd6f7666-6cc8-478d-9acf-21cfbad541c9" containerName="rabbitmq" Dec 10 19:24:36 crc kubenswrapper[4828]: E1210 19:24:36.867700 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b24f93de-d604-499b-be8e-2e934d119192" containerName="rabbitmq" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.867705 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="b24f93de-d604-499b-be8e-2e934d119192" containerName="rabbitmq" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.867929 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd6f7666-6cc8-478d-9acf-21cfbad541c9" containerName="rabbitmq" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.867948 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="b24f93de-d604-499b-be8e-2e934d119192" containerName="rabbitmq" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.869305 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.873505 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.873676 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.873877 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.873995 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.875509 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-96rhn" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.876106 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.876572 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.885399 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.888681 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.893882 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-zttjq" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.894074 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.894085 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.894201 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.894246 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.894420 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.897196 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.899303 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.912480 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.973599 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3eb66f68-0443-4122-8d37-8000aefcbc3b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.973675 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/540f0868-0b12-44cc-806a-9d31d644da25-config-data\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.973698 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/540f0868-0b12-44cc-806a-9d31d644da25-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.973724 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/540f0868-0b12-44cc-806a-9d31d644da25-pod-info\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.973757 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcbfl\" (UniqueName: \"kubernetes.io/projected/540f0868-0b12-44cc-806a-9d31d644da25-kube-api-access-hcbfl\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.973786 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3eb66f68-0443-4122-8d37-8000aefcbc3b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.973839 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/540f0868-0b12-44cc-806a-9d31d644da25-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.973869 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3eb66f68-0443-4122-8d37-8000aefcbc3b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.973887 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3eb66f68-0443-4122-8d37-8000aefcbc3b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.973910 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/540f0868-0b12-44cc-806a-9d31d644da25-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.973930 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/540f0868-0b12-44cc-806a-9d31d644da25-server-conf\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.973951 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3eb66f68-0443-4122-8d37-8000aefcbc3b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.974347 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.974435 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzcd6\" (UniqueName: \"kubernetes.io/projected/3eb66f68-0443-4122-8d37-8000aefcbc3b-kube-api-access-xzcd6\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.974502 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/540f0868-0b12-44cc-806a-9d31d644da25-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.974563 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3eb66f68-0443-4122-8d37-8000aefcbc3b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.974615 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/540f0868-0b12-44cc-806a-9d31d644da25-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.974788 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.974868 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/540f0868-0b12-44cc-806a-9d31d644da25-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.974930 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3eb66f68-0443-4122-8d37-8000aefcbc3b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.974992 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3eb66f68-0443-4122-8d37-8000aefcbc3b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:36 crc kubenswrapper[4828]: I1210 19:24:36.975039 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3eb66f68-0443-4122-8d37-8000aefcbc3b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077114 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/540f0868-0b12-44cc-806a-9d31d644da25-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077165 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3eb66f68-0443-4122-8d37-8000aefcbc3b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077192 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/540f0868-0b12-44cc-806a-9d31d644da25-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077219 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077248 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/540f0868-0b12-44cc-806a-9d31d644da25-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077277 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3eb66f68-0443-4122-8d37-8000aefcbc3b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077304 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3eb66f68-0443-4122-8d37-8000aefcbc3b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077332 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3eb66f68-0443-4122-8d37-8000aefcbc3b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077372 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3eb66f68-0443-4122-8d37-8000aefcbc3b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077411 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/540f0868-0b12-44cc-806a-9d31d644da25-config-data\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077427 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/540f0868-0b12-44cc-806a-9d31d644da25-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077449 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/540f0868-0b12-44cc-806a-9d31d644da25-pod-info\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077477 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcbfl\" (UniqueName: \"kubernetes.io/projected/540f0868-0b12-44cc-806a-9d31d644da25-kube-api-access-hcbfl\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077501 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3eb66f68-0443-4122-8d37-8000aefcbc3b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077533 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/540f0868-0b12-44cc-806a-9d31d644da25-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077560 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3eb66f68-0443-4122-8d37-8000aefcbc3b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077578 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3eb66f68-0443-4122-8d37-8000aefcbc3b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077591 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/540f0868-0b12-44cc-806a-9d31d644da25-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077609 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/540f0868-0b12-44cc-806a-9d31d644da25-server-conf\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077631 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3eb66f68-0443-4122-8d37-8000aefcbc3b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077664 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.077684 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzcd6\" (UniqueName: \"kubernetes.io/projected/3eb66f68-0443-4122-8d37-8000aefcbc3b-kube-api-access-xzcd6\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.079227 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/540f0868-0b12-44cc-806a-9d31d644da25-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.079444 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3eb66f68-0443-4122-8d37-8000aefcbc3b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.079535 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/540f0868-0b12-44cc-806a-9d31d644da25-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.082704 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3eb66f68-0443-4122-8d37-8000aefcbc3b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.083279 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3eb66f68-0443-4122-8d37-8000aefcbc3b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.090932 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/540f0868-0b12-44cc-806a-9d31d644da25-config-data\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.091327 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3eb66f68-0443-4122-8d37-8000aefcbc3b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.091366 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.091483 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3eb66f68-0443-4122-8d37-8000aefcbc3b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.092110 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/540f0868-0b12-44cc-806a-9d31d644da25-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.092202 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.097304 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/540f0868-0b12-44cc-806a-9d31d644da25-server-conf\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.146047 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3eb66f68-0443-4122-8d37-8000aefcbc3b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.146494 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3eb66f68-0443-4122-8d37-8000aefcbc3b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.146895 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/540f0868-0b12-44cc-806a-9d31d644da25-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.147299 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/540f0868-0b12-44cc-806a-9d31d644da25-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.147434 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzcd6\" (UniqueName: \"kubernetes.io/projected/3eb66f68-0443-4122-8d37-8000aefcbc3b-kube-api-access-xzcd6\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.147897 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/540f0868-0b12-44cc-806a-9d31d644da25-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.148419 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3eb66f68-0443-4122-8d37-8000aefcbc3b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.164538 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/540f0868-0b12-44cc-806a-9d31d644da25-pod-info\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.165843 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcbfl\" (UniqueName: \"kubernetes.io/projected/540f0868-0b12-44cc-806a-9d31d644da25-kube-api-access-hcbfl\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.166354 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3eb66f68-0443-4122-8d37-8000aefcbc3b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.339469 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3eb66f68-0443-4122-8d37-8000aefcbc3b\") " pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.352240 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"540f0868-0b12-44cc-806a-9d31d644da25\") " pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.497897 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.568402 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="fd6f7666-6cc8-478d-9acf-21cfbad541c9" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.129:5671: i/o timeout" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.582415 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.655908 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="b24f93de-d604-499b-be8e-2e934d119192" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.130:5671: i/o timeout" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.790541 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:24:37 crc kubenswrapper[4828]: E1210 19:24:37.790772 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.805159 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b24f93de-d604-499b-be8e-2e934d119192" path="/var/lib/kubelet/pods/b24f93de-d604-499b-be8e-2e934d119192/volumes" Dec 10 19:24:37 crc kubenswrapper[4828]: I1210 19:24:37.806506 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd6f7666-6cc8-478d-9acf-21cfbad541c9" path="/var/lib/kubelet/pods/fd6f7666-6cc8-478d-9acf-21cfbad541c9/volumes" Dec 10 19:24:38 crc kubenswrapper[4828]: E1210 19:24:38.444501 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 10 19:24:38 crc kubenswrapper[4828]: E1210 19:24:38.444832 4828 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Dec 10 19:24:38 crc kubenswrapper[4828]: E1210 19:24:38.445333 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n88h65h56chc7h5bh667h5hb9h647h54dh7h697h665hb4h549hcfh5ffh697h66h698h585h87h6h5c7hd7hb7h674h554h59dh7h668h555q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qw4md,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(cd0c6bc2-1b33-4c82-ae2f-9ea265564b82): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 19:24:38 crc kubenswrapper[4828]: I1210 19:24:38.461424 4828 scope.go:117] "RemoveContainer" containerID="5505417dc4fa51ffbfa46576fe4ceb89ae8a05311568421b1979522bbc329385" Dec 10 19:24:38 crc kubenswrapper[4828]: I1210 19:24:38.561716 4828 scope.go:117] "RemoveContainer" containerID="3cf3999e422c09e05f979b2209c544abc8e36d082b76293e4209e83d892753ba" Dec 10 19:24:38 crc kubenswrapper[4828]: I1210 19:24:38.674993 4828 scope.go:117] "RemoveContainer" containerID="9fb4907035796b871225aeb7750d239070e57fcc3199b135138f56a1c472ce23" Dec 10 19:24:39 crc kubenswrapper[4828]: I1210 19:24:39.075349 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-k6pwv"] Dec 10 19:24:39 crc kubenswrapper[4828]: W1210 19:24:39.081131 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74792a2b_341a_4065_93e3_215c325db10c.slice/crio-a6aa9472f68905cf0f279458f3a1f3de427c2533a06080e6dc8a5eeff5b23dde WatchSource:0}: Error finding container a6aa9472f68905cf0f279458f3a1f3de427c2533a06080e6dc8a5eeff5b23dde: Status 404 returned error can't find the container with id a6aa9472f68905cf0f279458f3a1f3de427c2533a06080e6dc8a5eeff5b23dde Dec 10 19:24:39 crc kubenswrapper[4828]: I1210 19:24:39.086512 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 10 19:24:39 crc kubenswrapper[4828]: W1210 19:24:39.088390 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3eb66f68_0443_4122_8d37_8000aefcbc3b.slice/crio-3b605a0db970947037b7a9131cbdad563299214e68eff0060937820fd7244ad0 WatchSource:0}: Error finding container 3b605a0db970947037b7a9131cbdad563299214e68eff0060937820fd7244ad0: Status 404 returned error can't find the container with id 3b605a0db970947037b7a9131cbdad563299214e68eff0060937820fd7244ad0 Dec 10 19:24:39 crc kubenswrapper[4828]: I1210 19:24:39.239979 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 10 19:24:39 crc kubenswrapper[4828]: I1210 19:24:39.267300 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"540f0868-0b12-44cc-806a-9d31d644da25","Type":"ContainerStarted","Data":"5b01b0769e9a87663bdf18885e3b26564c332bc46f1e0dfe50861955d7dd8e8e"} Dec 10 19:24:39 crc kubenswrapper[4828]: I1210 19:24:39.269643 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82","Type":"ContainerStarted","Data":"7e44cb7b7b45af73f58adf291f466949baac839279cbc72f483bfdddbc224b02"} Dec 10 19:24:39 crc kubenswrapper[4828]: I1210 19:24:39.275548 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" event={"ID":"74792a2b-341a-4065-93e3-215c325db10c","Type":"ContainerStarted","Data":"a6aa9472f68905cf0f279458f3a1f3de427c2533a06080e6dc8a5eeff5b23dde"} Dec 10 19:24:39 crc kubenswrapper[4828]: I1210 19:24:39.277302 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3eb66f68-0443-4122-8d37-8000aefcbc3b","Type":"ContainerStarted","Data":"3b605a0db970947037b7a9131cbdad563299214e68eff0060937820fd7244ad0"} Dec 10 19:24:40 crc kubenswrapper[4828]: I1210 19:24:40.302876 4828 generic.go:334] "Generic (PLEG): container finished" podID="74792a2b-341a-4065-93e3-215c325db10c" containerID="8f5f1377dad6ef982d63e83d8849fe435180a62cf5b2650e456e971465502ae0" exitCode=0 Dec 10 19:24:40 crc kubenswrapper[4828]: I1210 19:24:40.303259 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" event={"ID":"74792a2b-341a-4065-93e3-215c325db10c","Type":"ContainerDied","Data":"8f5f1377dad6ef982d63e83d8849fe435180a62cf5b2650e456e971465502ae0"} Dec 10 19:24:41 crc kubenswrapper[4828]: I1210 19:24:41.317345 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82","Type":"ContainerStarted","Data":"3c39dbeaf2f8ab989e4aea02df96b217fd9f0dbb9ded0d8585aa75eca78fe0e6"} Dec 10 19:24:41 crc kubenswrapper[4828]: I1210 19:24:41.320068 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" event={"ID":"74792a2b-341a-4065-93e3-215c325db10c","Type":"ContainerStarted","Data":"db3ac48041235c638079fa2fe5577eb248918091b89d8da15c3b99b484b075b9"} Dec 10 19:24:41 crc kubenswrapper[4828]: I1210 19:24:41.321106 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:41 crc kubenswrapper[4828]: I1210 19:24:41.323578 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3eb66f68-0443-4122-8d37-8000aefcbc3b","Type":"ContainerStarted","Data":"81b437b4051791af4f516370cff7655f490ad68da013321b50f0b02df1c4abc1"} Dec 10 19:24:41 crc kubenswrapper[4828]: I1210 19:24:41.342744 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" podStartSLOduration=13.342727364 podStartE2EDuration="13.342727364s" podCreationTimestamp="2025-12-10 19:24:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:24:41.338129462 +0000 UTC m=+1761.848740467" watchObservedRunningTime="2025-12-10 19:24:41.342727364 +0000 UTC m=+1761.853338369" Dec 10 19:24:42 crc kubenswrapper[4828]: E1210 19:24:42.235544 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="cd0c6bc2-1b33-4c82-ae2f-9ea265564b82" Dec 10 19:24:42 crc kubenswrapper[4828]: I1210 19:24:42.337600 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82","Type":"ContainerStarted","Data":"b6d96f0b9d4441d361a036281eb6a3b4fea6623d586eac19182c3564d6f3ad5a"} Dec 10 19:24:42 crc kubenswrapper[4828]: I1210 19:24:42.337736 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 10 19:24:42 crc kubenswrapper[4828]: E1210 19:24:42.339086 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="cd0c6bc2-1b33-4c82-ae2f-9ea265564b82" Dec 10 19:24:42 crc kubenswrapper[4828]: I1210 19:24:42.339405 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"540f0868-0b12-44cc-806a-9d31d644da25","Type":"ContainerStarted","Data":"bb6e22e205bb8ee4d62d7b1021e1ea28e7d87c0650c394c1fa096feafbbd5683"} Dec 10 19:24:43 crc kubenswrapper[4828]: E1210 19:24:43.386130 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="cd0c6bc2-1b33-4c82-ae2f-9ea265564b82" Dec 10 19:24:48 crc kubenswrapper[4828]: I1210 19:24:48.937004 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.019027 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-ncgpx"] Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.019598 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" podUID="0c16730a-9ba1-48c7-93f8-f10980ca44ab" containerName="dnsmasq-dns" containerID="cri-o://7cceaf33cfacf6372e9726e3cbed0ef4b721b062ce3a3307e4305f4313c61666" gracePeriod=10 Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.182427 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d75f767dc-9xsbz"] Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.185430 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.211672 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d75f767dc-9xsbz"] Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.309116 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.310183 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-config\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.310417 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-dns-swift-storage-0\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.310596 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-ovsdbserver-nb\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.310732 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-ovsdbserver-sb\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.310839 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-dns-svc\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.310930 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbkqp\" (UniqueName: \"kubernetes.io/projected/b460f485-38dd-48c1-81d5-288cd81f392b-kube-api-access-nbkqp\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.413668 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-config\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.413773 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-dns-swift-storage-0\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.413829 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-ovsdbserver-nb\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.413870 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-ovsdbserver-sb\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.413897 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-dns-svc\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.413926 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbkqp\" (UniqueName: \"kubernetes.io/projected/b460f485-38dd-48c1-81d5-288cd81f392b-kube-api-access-nbkqp\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.413948 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.415026 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.416110 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-config\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.416515 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-ovsdbserver-sb\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.417151 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-dns-swift-storage-0\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.417170 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-ovsdbserver-nb\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.417583 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b460f485-38dd-48c1-81d5-288cd81f392b-dns-svc\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.437252 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbkqp\" (UniqueName: \"kubernetes.io/projected/b460f485-38dd-48c1-81d5-288cd81f392b-kube-api-access-nbkqp\") pod \"dnsmasq-dns-5d75f767dc-9xsbz\" (UID: \"b460f485-38dd-48c1-81d5-288cd81f392b\") " pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.443214 4828 generic.go:334] "Generic (PLEG): container finished" podID="0c16730a-9ba1-48c7-93f8-f10980ca44ab" containerID="7cceaf33cfacf6372e9726e3cbed0ef4b721b062ce3a3307e4305f4313c61666" exitCode=0 Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.443262 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" event={"ID":"0c16730a-9ba1-48c7-93f8-f10980ca44ab","Type":"ContainerDied","Data":"7cceaf33cfacf6372e9726e3cbed0ef4b721b062ce3a3307e4305f4313c61666"} Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.523297 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.735378 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.822241 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-dns-svc\") pod \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.822314 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-ovsdbserver-nb\") pod \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.822501 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-dns-swift-storage-0\") pod \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.822522 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7nqd\" (UniqueName: \"kubernetes.io/projected/0c16730a-9ba1-48c7-93f8-f10980ca44ab-kube-api-access-w7nqd\") pod \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.822547 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-config\") pod \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.822701 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-ovsdbserver-sb\") pod \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\" (UID: \"0c16730a-9ba1-48c7-93f8-f10980ca44ab\") " Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.892454 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c16730a-9ba1-48c7-93f8-f10980ca44ab-kube-api-access-w7nqd" (OuterVolumeSpecName: "kube-api-access-w7nqd") pod "0c16730a-9ba1-48c7-93f8-f10980ca44ab" (UID: "0c16730a-9ba1-48c7-93f8-f10980ca44ab"). InnerVolumeSpecName "kube-api-access-w7nqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:24:49 crc kubenswrapper[4828]: I1210 19:24:49.935046 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7nqd\" (UniqueName: \"kubernetes.io/projected/0c16730a-9ba1-48c7-93f8-f10980ca44ab-kube-api-access-w7nqd\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.012174 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0c16730a-9ba1-48c7-93f8-f10980ca44ab" (UID: "0c16730a-9ba1-48c7-93f8-f10980ca44ab"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.070362 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.108135 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-config" (OuterVolumeSpecName: "config") pod "0c16730a-9ba1-48c7-93f8-f10980ca44ab" (UID: "0c16730a-9ba1-48c7-93f8-f10980ca44ab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.116612 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0c16730a-9ba1-48c7-93f8-f10980ca44ab" (UID: "0c16730a-9ba1-48c7-93f8-f10980ca44ab"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.123545 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0c16730a-9ba1-48c7-93f8-f10980ca44ab" (UID: "0c16730a-9ba1-48c7-93f8-f10980ca44ab"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.179560 4828 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.179603 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.179616 4828 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.199551 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d75f767dc-9xsbz"] Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.201706 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0c16730a-9ba1-48c7-93f8-f10980ca44ab" (UID: "0c16730a-9ba1-48c7-93f8-f10980ca44ab"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.283927 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c16730a-9ba1-48c7-93f8-f10980ca44ab-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.458569 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" event={"ID":"0c16730a-9ba1-48c7-93f8-f10980ca44ab","Type":"ContainerDied","Data":"6adaf9f1012dc20dcceac3b0234eab63f3c3bcd8a0c8c6018903402a3a80b7cf"} Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.458629 4828 scope.go:117] "RemoveContainer" containerID="7cceaf33cfacf6372e9726e3cbed0ef4b721b062ce3a3307e4305f4313c61666" Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.458775 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-ncgpx" Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.461178 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" event={"ID":"b460f485-38dd-48c1-81d5-288cd81f392b","Type":"ContainerStarted","Data":"9c0e31d1196833a81ebe96fe5d7e06e36a6833c277cd17b03ea95aba763b1eb2"} Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.486839 4828 scope.go:117] "RemoveContainer" containerID="6b91b4ce91d9e2ad35618daa9cc5cf8f27497a876bd80f937b05a985356835ad" Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.503416 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-ncgpx"] Dec 10 19:24:50 crc kubenswrapper[4828]: I1210 19:24:50.516902 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-ncgpx"] Dec 10 19:24:51 crc kubenswrapper[4828]: I1210 19:24:51.473013 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2n2fz" event={"ID":"2449558a-d727-461c-acb2-99e380868fb1","Type":"ContainerStarted","Data":"2ac69f9a7894f1dd0075acf25f01223ed2fcb8080f0630897935c14a98d127ab"} Dec 10 19:24:51 crc kubenswrapper[4828]: I1210 19:24:51.475783 4828 generic.go:334] "Generic (PLEG): container finished" podID="b460f485-38dd-48c1-81d5-288cd81f392b" containerID="775a3db38755454401d26541bbf7311f3e5d963a33877b57a14ff838d89d66b1" exitCode=0 Dec 10 19:24:51 crc kubenswrapper[4828]: I1210 19:24:51.475858 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" event={"ID":"b460f485-38dd-48c1-81d5-288cd81f392b","Type":"ContainerDied","Data":"775a3db38755454401d26541bbf7311f3e5d963a33877b57a14ff838d89d66b1"} Dec 10 19:24:51 crc kubenswrapper[4828]: I1210 19:24:51.496177 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-2n2fz" podStartSLOduration=2.502226413 podStartE2EDuration="40.496155999s" podCreationTimestamp="2025-12-10 19:24:11 +0000 UTC" firstStartedPulling="2025-12-10 19:24:12.119070261 +0000 UTC m=+1732.629681266" lastFinishedPulling="2025-12-10 19:24:50.112999857 +0000 UTC m=+1770.623610852" observedRunningTime="2025-12-10 19:24:51.490643133 +0000 UTC m=+1772.001254138" watchObservedRunningTime="2025-12-10 19:24:51.496155999 +0000 UTC m=+1772.006767024" Dec 10 19:24:51 crc kubenswrapper[4828]: I1210 19:24:51.801950 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c16730a-9ba1-48c7-93f8-f10980ca44ab" path="/var/lib/kubelet/pods/0c16730a-9ba1-48c7-93f8-f10980ca44ab/volumes" Dec 10 19:24:52 crc kubenswrapper[4828]: I1210 19:24:52.491561 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" event={"ID":"b460f485-38dd-48c1-81d5-288cd81f392b","Type":"ContainerStarted","Data":"61ba38de689ed28d0199799862c609ee3249220abaf1ac43e339e7de401be512"} Dec 10 19:24:52 crc kubenswrapper[4828]: I1210 19:24:52.491701 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:52 crc kubenswrapper[4828]: I1210 19:24:52.509375 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" podStartSLOduration=3.509354512 podStartE2EDuration="3.509354512s" podCreationTimestamp="2025-12-10 19:24:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:24:52.506710642 +0000 UTC m=+1773.017321657" watchObservedRunningTime="2025-12-10 19:24:52.509354512 +0000 UTC m=+1773.019965517" Dec 10 19:24:52 crc kubenswrapper[4828]: I1210 19:24:52.789432 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:24:52 crc kubenswrapper[4828]: E1210 19:24:52.790378 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:24:53 crc kubenswrapper[4828]: I1210 19:24:53.507934 4828 generic.go:334] "Generic (PLEG): container finished" podID="2449558a-d727-461c-acb2-99e380868fb1" containerID="2ac69f9a7894f1dd0075acf25f01223ed2fcb8080f0630897935c14a98d127ab" exitCode=0 Dec 10 19:24:53 crc kubenswrapper[4828]: I1210 19:24:53.510726 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2n2fz" event={"ID":"2449558a-d727-461c-acb2-99e380868fb1","Type":"ContainerDied","Data":"2ac69f9a7894f1dd0075acf25f01223ed2fcb8080f0630897935c14a98d127ab"} Dec 10 19:24:54 crc kubenswrapper[4828]: I1210 19:24:54.940291 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2n2fz" Dec 10 19:24:54 crc kubenswrapper[4828]: I1210 19:24:54.999046 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2449558a-d727-461c-acb2-99e380868fb1-combined-ca-bundle\") pod \"2449558a-d727-461c-acb2-99e380868fb1\" (UID: \"2449558a-d727-461c-acb2-99e380868fb1\") " Dec 10 19:24:54 crc kubenswrapper[4828]: I1210 19:24:54.999167 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvsmv\" (UniqueName: \"kubernetes.io/projected/2449558a-d727-461c-acb2-99e380868fb1-kube-api-access-vvsmv\") pod \"2449558a-d727-461c-acb2-99e380868fb1\" (UID: \"2449558a-d727-461c-acb2-99e380868fb1\") " Dec 10 19:24:54 crc kubenswrapper[4828]: I1210 19:24:54.999233 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2449558a-d727-461c-acb2-99e380868fb1-config-data\") pod \"2449558a-d727-461c-acb2-99e380868fb1\" (UID: \"2449558a-d727-461c-acb2-99e380868fb1\") " Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.005003 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2449558a-d727-461c-acb2-99e380868fb1-kube-api-access-vvsmv" (OuterVolumeSpecName: "kube-api-access-vvsmv") pod "2449558a-d727-461c-acb2-99e380868fb1" (UID: "2449558a-d727-461c-acb2-99e380868fb1"). InnerVolumeSpecName "kube-api-access-vvsmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.033287 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2449558a-d727-461c-acb2-99e380868fb1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2449558a-d727-461c-acb2-99e380868fb1" (UID: "2449558a-d727-461c-acb2-99e380868fb1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.085502 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2449558a-d727-461c-acb2-99e380868fb1-config-data" (OuterVolumeSpecName: "config-data") pod "2449558a-d727-461c-acb2-99e380868fb1" (UID: "2449558a-d727-461c-acb2-99e380868fb1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.107431 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2449558a-d727-461c-acb2-99e380868fb1-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.107471 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2449558a-d727-461c-acb2-99e380868fb1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.107482 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvsmv\" (UniqueName: \"kubernetes.io/projected/2449558a-d727-461c-acb2-99e380868fb1-kube-api-access-vvsmv\") on node \"crc\" DevicePath \"\"" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.144749 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j2hgw"] Dec 10 19:24:55 crc kubenswrapper[4828]: E1210 19:24:55.145471 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c16730a-9ba1-48c7-93f8-f10980ca44ab" containerName="init" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.145495 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c16730a-9ba1-48c7-93f8-f10980ca44ab" containerName="init" Dec 10 19:24:55 crc kubenswrapper[4828]: E1210 19:24:55.145525 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2449558a-d727-461c-acb2-99e380868fb1" containerName="heat-db-sync" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.145534 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="2449558a-d727-461c-acb2-99e380868fb1" containerName="heat-db-sync" Dec 10 19:24:55 crc kubenswrapper[4828]: E1210 19:24:55.145555 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c16730a-9ba1-48c7-93f8-f10980ca44ab" containerName="dnsmasq-dns" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.145564 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c16730a-9ba1-48c7-93f8-f10980ca44ab" containerName="dnsmasq-dns" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.145880 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="2449558a-d727-461c-acb2-99e380868fb1" containerName="heat-db-sync" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.145912 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c16730a-9ba1-48c7-93f8-f10980ca44ab" containerName="dnsmasq-dns" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.148091 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j2hgw" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.158922 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j2hgw"] Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.209100 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-utilities\") pod \"community-operators-j2hgw\" (UID: \"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097\") " pod="openshift-marketplace/community-operators-j2hgw" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.209197 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fzt6\" (UniqueName: \"kubernetes.io/projected/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-kube-api-access-8fzt6\") pod \"community-operators-j2hgw\" (UID: \"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097\") " pod="openshift-marketplace/community-operators-j2hgw" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.209266 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-catalog-content\") pod \"community-operators-j2hgw\" (UID: \"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097\") " pod="openshift-marketplace/community-operators-j2hgw" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.311960 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-utilities\") pod \"community-operators-j2hgw\" (UID: \"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097\") " pod="openshift-marketplace/community-operators-j2hgw" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.312062 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fzt6\" (UniqueName: \"kubernetes.io/projected/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-kube-api-access-8fzt6\") pod \"community-operators-j2hgw\" (UID: \"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097\") " pod="openshift-marketplace/community-operators-j2hgw" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.312136 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-catalog-content\") pod \"community-operators-j2hgw\" (UID: \"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097\") " pod="openshift-marketplace/community-operators-j2hgw" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.312614 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-utilities\") pod \"community-operators-j2hgw\" (UID: \"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097\") " pod="openshift-marketplace/community-operators-j2hgw" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.313066 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-catalog-content\") pod \"community-operators-j2hgw\" (UID: \"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097\") " pod="openshift-marketplace/community-operators-j2hgw" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.329518 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fzt6\" (UniqueName: \"kubernetes.io/projected/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-kube-api-access-8fzt6\") pod \"community-operators-j2hgw\" (UID: \"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097\") " pod="openshift-marketplace/community-operators-j2hgw" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.501982 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j2hgw" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.561105 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2n2fz" event={"ID":"2449558a-d727-461c-acb2-99e380868fb1","Type":"ContainerDied","Data":"fe3b73df95e072c9b7023f992c751176eb0e8b0d2d9f21d6a81247fddbef3cc0"} Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.561147 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe3b73df95e072c9b7023f992c751176eb0e8b0d2d9f21d6a81247fddbef3cc0" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.561218 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2n2fz" Dec 10 19:24:55 crc kubenswrapper[4828]: I1210 19:24:55.807191 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 10 19:24:56 crc kubenswrapper[4828]: W1210 19:24:56.084624 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca9f07ac_e1bc_42d2_9f45_c3b4b3843097.slice/crio-342e5640c0465e251d7c8245913f75bf07c1c6eebe6d71a4bd5c7c899e0f4e93 WatchSource:0}: Error finding container 342e5640c0465e251d7c8245913f75bf07c1c6eebe6d71a4bd5c7c899e0f4e93: Status 404 returned error can't find the container with id 342e5640c0465e251d7c8245913f75bf07c1c6eebe6d71a4bd5c7c899e0f4e93 Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.089478 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j2hgw"] Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.553785 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-55d5c67ddc-vcr2w"] Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.557925 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-55d5c67ddc-vcr2w" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.572197 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-55d5c67ddc-vcr2w"] Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.614110 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cd0c6bc2-1b33-4c82-ae2f-9ea265564b82","Type":"ContainerStarted","Data":"ecdf725381153d75e94a92f5181086f983210c8809cecb2ad45572b155a4dfa2"} Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.649351 4828 generic.go:334] "Generic (PLEG): container finished" podID="ca9f07ac-e1bc-42d2-9f45-c3b4b3843097" containerID="0056fff6b7bc9d122721ca99f1581024ab0338687fec8f4aa7dd1618525d7019" exitCode=0 Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.649412 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2hgw" event={"ID":"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097","Type":"ContainerDied","Data":"0056fff6b7bc9d122721ca99f1581024ab0338687fec8f4aa7dd1618525d7019"} Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.649456 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2hgw" event={"ID":"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097","Type":"ContainerStarted","Data":"342e5640c0465e251d7c8245913f75bf07c1c6eebe6d71a4bd5c7c899e0f4e93"} Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.651577 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd195e2-5f2e-4472-a481-bfb485dcfc74-config-data\") pod \"heat-engine-55d5c67ddc-vcr2w\" (UID: \"7dd195e2-5f2e-4472-a481-bfb485dcfc74\") " pod="openstack/heat-engine-55d5c67ddc-vcr2w" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.651982 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd195e2-5f2e-4472-a481-bfb485dcfc74-combined-ca-bundle\") pod \"heat-engine-55d5c67ddc-vcr2w\" (UID: \"7dd195e2-5f2e-4472-a481-bfb485dcfc74\") " pod="openstack/heat-engine-55d5c67ddc-vcr2w" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.672164 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfr4q\" (UniqueName: \"kubernetes.io/projected/7dd195e2-5f2e-4472-a481-bfb485dcfc74-kube-api-access-xfr4q\") pod \"heat-engine-55d5c67ddc-vcr2w\" (UID: \"7dd195e2-5f2e-4472-a481-bfb485dcfc74\") " pod="openstack/heat-engine-55d5c67ddc-vcr2w" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.672440 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7dd195e2-5f2e-4472-a481-bfb485dcfc74-config-data-custom\") pod \"heat-engine-55d5c67ddc-vcr2w\" (UID: \"7dd195e2-5f2e-4472-a481-bfb485dcfc74\") " pod="openstack/heat-engine-55d5c67ddc-vcr2w" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.672881 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5859bd5774-r9zl7"] Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.674577 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.708386 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5859bd5774-r9zl7"] Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.724389 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-cb5d96886-n5cp9"] Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.726258 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.816020135 podStartE2EDuration="41.726238273s" podCreationTimestamp="2025-12-10 19:24:15 +0000 UTC" firstStartedPulling="2025-12-10 19:24:16.111826915 +0000 UTC m=+1736.622437920" lastFinishedPulling="2025-12-10 19:24:56.022045053 +0000 UTC m=+1776.532656058" observedRunningTime="2025-12-10 19:24:56.651935745 +0000 UTC m=+1777.162546770" watchObservedRunningTime="2025-12-10 19:24:56.726238273 +0000 UTC m=+1777.236849268" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.727250 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.764924 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-cb5d96886-n5cp9"] Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.775860 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd195e2-5f2e-4472-a481-bfb485dcfc74-config-data\") pod \"heat-engine-55d5c67ddc-vcr2w\" (UID: \"7dd195e2-5f2e-4472-a481-bfb485dcfc74\") " pod="openstack/heat-engine-55d5c67ddc-vcr2w" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.775940 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2279x\" (UniqueName: \"kubernetes.io/projected/961e506a-98a1-48e1-afd5-c4326af0bdb3-kube-api-access-2279x\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.775980 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd195e2-5f2e-4472-a481-bfb485dcfc74-combined-ca-bundle\") pod \"heat-engine-55d5c67ddc-vcr2w\" (UID: \"7dd195e2-5f2e-4472-a481-bfb485dcfc74\") " pod="openstack/heat-engine-55d5c67ddc-vcr2w" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.776046 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/961e506a-98a1-48e1-afd5-c4326af0bdb3-internal-tls-certs\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.776137 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/961e506a-98a1-48e1-afd5-c4326af0bdb3-config-data\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.776171 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfr4q\" (UniqueName: \"kubernetes.io/projected/7dd195e2-5f2e-4472-a481-bfb485dcfc74-kube-api-access-xfr4q\") pod \"heat-engine-55d5c67ddc-vcr2w\" (UID: \"7dd195e2-5f2e-4472-a481-bfb485dcfc74\") " pod="openstack/heat-engine-55d5c67ddc-vcr2w" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.776235 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/961e506a-98a1-48e1-afd5-c4326af0bdb3-combined-ca-bundle\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.776264 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/961e506a-98a1-48e1-afd5-c4326af0bdb3-public-tls-certs\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.776355 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7dd195e2-5f2e-4472-a481-bfb485dcfc74-config-data-custom\") pod \"heat-engine-55d5c67ddc-vcr2w\" (UID: \"7dd195e2-5f2e-4472-a481-bfb485dcfc74\") " pod="openstack/heat-engine-55d5c67ddc-vcr2w" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.776579 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/961e506a-98a1-48e1-afd5-c4326af0bdb3-config-data-custom\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.786080 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd195e2-5f2e-4472-a481-bfb485dcfc74-combined-ca-bundle\") pod \"heat-engine-55d5c67ddc-vcr2w\" (UID: \"7dd195e2-5f2e-4472-a481-bfb485dcfc74\") " pod="openstack/heat-engine-55d5c67ddc-vcr2w" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.786371 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7dd195e2-5f2e-4472-a481-bfb485dcfc74-config-data-custom\") pod \"heat-engine-55d5c67ddc-vcr2w\" (UID: \"7dd195e2-5f2e-4472-a481-bfb485dcfc74\") " pod="openstack/heat-engine-55d5c67ddc-vcr2w" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.786896 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd195e2-5f2e-4472-a481-bfb485dcfc74-config-data\") pod \"heat-engine-55d5c67ddc-vcr2w\" (UID: \"7dd195e2-5f2e-4472-a481-bfb485dcfc74\") " pod="openstack/heat-engine-55d5c67ddc-vcr2w" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.803404 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfr4q\" (UniqueName: \"kubernetes.io/projected/7dd195e2-5f2e-4472-a481-bfb485dcfc74-kube-api-access-xfr4q\") pod \"heat-engine-55d5c67ddc-vcr2w\" (UID: \"7dd195e2-5f2e-4472-a481-bfb485dcfc74\") " pod="openstack/heat-engine-55d5c67ddc-vcr2w" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.878950 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1887db4-171d-4bd1-8337-4a8c72e9e93d-config-data-custom\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.879016 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/961e506a-98a1-48e1-afd5-c4326af0bdb3-config-data-custom\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.879060 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1887db4-171d-4bd1-8337-4a8c72e9e93d-combined-ca-bundle\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.879121 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2279x\" (UniqueName: \"kubernetes.io/projected/961e506a-98a1-48e1-afd5-c4326af0bdb3-kube-api-access-2279x\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.879308 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/961e506a-98a1-48e1-afd5-c4326af0bdb3-internal-tls-certs\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.879489 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/961e506a-98a1-48e1-afd5-c4326af0bdb3-config-data\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.879629 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1887db4-171d-4bd1-8337-4a8c72e9e93d-internal-tls-certs\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.879660 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1887db4-171d-4bd1-8337-4a8c72e9e93d-public-tls-certs\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.879684 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/961e506a-98a1-48e1-afd5-c4326af0bdb3-combined-ca-bundle\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.879703 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/961e506a-98a1-48e1-afd5-c4326af0bdb3-public-tls-certs\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.879847 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh59b\" (UniqueName: \"kubernetes.io/projected/d1887db4-171d-4bd1-8337-4a8c72e9e93d-kube-api-access-rh59b\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.879891 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1887db4-171d-4bd1-8337-4a8c72e9e93d-config-data\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.883486 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/961e506a-98a1-48e1-afd5-c4326af0bdb3-config-data-custom\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.884695 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/961e506a-98a1-48e1-afd5-c4326af0bdb3-internal-tls-certs\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.887580 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/961e506a-98a1-48e1-afd5-c4326af0bdb3-combined-ca-bundle\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.891145 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/961e506a-98a1-48e1-afd5-c4326af0bdb3-public-tls-certs\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.891571 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/961e506a-98a1-48e1-afd5-c4326af0bdb3-config-data\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.895684 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2279x\" (UniqueName: \"kubernetes.io/projected/961e506a-98a1-48e1-afd5-c4326af0bdb3-kube-api-access-2279x\") pod \"heat-api-5859bd5774-r9zl7\" (UID: \"961e506a-98a1-48e1-afd5-c4326af0bdb3\") " pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.932275 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-55d5c67ddc-vcr2w" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.981621 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1887db4-171d-4bd1-8337-4a8c72e9e93d-internal-tls-certs\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.981658 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1887db4-171d-4bd1-8337-4a8c72e9e93d-public-tls-certs\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.981712 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh59b\" (UniqueName: \"kubernetes.io/projected/d1887db4-171d-4bd1-8337-4a8c72e9e93d-kube-api-access-rh59b\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.981739 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1887db4-171d-4bd1-8337-4a8c72e9e93d-config-data\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.981871 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1887db4-171d-4bd1-8337-4a8c72e9e93d-config-data-custom\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.983846 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1887db4-171d-4bd1-8337-4a8c72e9e93d-combined-ca-bundle\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.985718 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1887db4-171d-4bd1-8337-4a8c72e9e93d-internal-tls-certs\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.992400 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1887db4-171d-4bd1-8337-4a8c72e9e93d-config-data\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.995186 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1887db4-171d-4bd1-8337-4a8c72e9e93d-combined-ca-bundle\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.996897 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh59b\" (UniqueName: \"kubernetes.io/projected/d1887db4-171d-4bd1-8337-4a8c72e9e93d-kube-api-access-rh59b\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.997499 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d1887db4-171d-4bd1-8337-4a8c72e9e93d-config-data-custom\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:56 crc kubenswrapper[4828]: I1210 19:24:56.998112 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d1887db4-171d-4bd1-8337-4a8c72e9e93d-public-tls-certs\") pod \"heat-cfnapi-cb5d96886-n5cp9\" (UID: \"d1887db4-171d-4bd1-8337-4a8c72e9e93d\") " pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:57 crc kubenswrapper[4828]: I1210 19:24:57.023089 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:24:57 crc kubenswrapper[4828]: I1210 19:24:57.067378 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:24:57 crc kubenswrapper[4828]: I1210 19:24:57.579101 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-55d5c67ddc-vcr2w"] Dec 10 19:24:57 crc kubenswrapper[4828]: W1210 19:24:57.590029 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod961e506a_98a1_48e1_afd5_c4326af0bdb3.slice/crio-e477329ccf813aa45939f1da7c0b44a138485fbfaaf913bb70565533bbd87e9a WatchSource:0}: Error finding container e477329ccf813aa45939f1da7c0b44a138485fbfaaf913bb70565533bbd87e9a: Status 404 returned error can't find the container with id e477329ccf813aa45939f1da7c0b44a138485fbfaaf913bb70565533bbd87e9a Dec 10 19:24:57 crc kubenswrapper[4828]: I1210 19:24:57.598997 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5859bd5774-r9zl7"] Dec 10 19:24:57 crc kubenswrapper[4828]: I1210 19:24:57.666297 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5859bd5774-r9zl7" event={"ID":"961e506a-98a1-48e1-afd5-c4326af0bdb3","Type":"ContainerStarted","Data":"e477329ccf813aa45939f1da7c0b44a138485fbfaaf913bb70565533bbd87e9a"} Dec 10 19:24:57 crc kubenswrapper[4828]: I1210 19:24:57.669141 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2hgw" event={"ID":"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097","Type":"ContainerStarted","Data":"73625e2e7afed03b8010a8c1eca56d7ffb106f26c1523e5fa675968974984262"} Dec 10 19:24:57 crc kubenswrapper[4828]: I1210 19:24:57.672592 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-55d5c67ddc-vcr2w" event={"ID":"7dd195e2-5f2e-4472-a481-bfb485dcfc74","Type":"ContainerStarted","Data":"6dfbf9c12f72263face80fb66875d8392c51d61505d84e47d95fb343b9fe51d6"} Dec 10 19:24:57 crc kubenswrapper[4828]: I1210 19:24:57.774401 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-cb5d96886-n5cp9"] Dec 10 19:24:58 crc kubenswrapper[4828]: I1210 19:24:58.684332 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-55d5c67ddc-vcr2w" event={"ID":"7dd195e2-5f2e-4472-a481-bfb485dcfc74","Type":"ContainerStarted","Data":"72374c5f2fc4c0052bae61b15ca8a92df70688e6879b675d438b26c7a40f05af"} Dec 10 19:24:58 crc kubenswrapper[4828]: I1210 19:24:58.684672 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-55d5c67ddc-vcr2w" Dec 10 19:24:58 crc kubenswrapper[4828]: I1210 19:24:58.686565 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cb5d96886-n5cp9" event={"ID":"d1887db4-171d-4bd1-8337-4a8c72e9e93d","Type":"ContainerStarted","Data":"00edb4103a00fcbe04d58b7a99630febc99527eae7db8787019164384f5aed1f"} Dec 10 19:24:58 crc kubenswrapper[4828]: I1210 19:24:58.729926 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-55d5c67ddc-vcr2w" podStartSLOduration=2.729761206 podStartE2EDuration="2.729761206s" podCreationTimestamp="2025-12-10 19:24:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:24:58.706789528 +0000 UTC m=+1779.217400523" watchObservedRunningTime="2025-12-10 19:24:58.729761206 +0000 UTC m=+1779.240372211" Dec 10 19:24:59 crc kubenswrapper[4828]: I1210 19:24:59.524945 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d75f767dc-9xsbz" Dec 10 19:24:59 crc kubenswrapper[4828]: I1210 19:24:59.594936 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-k6pwv"] Dec 10 19:24:59 crc kubenswrapper[4828]: I1210 19:24:59.595164 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" podUID="74792a2b-341a-4065-93e3-215c325db10c" containerName="dnsmasq-dns" containerID="cri-o://db3ac48041235c638079fa2fe5577eb248918091b89d8da15c3b99b484b075b9" gracePeriod=10 Dec 10 19:25:01 crc kubenswrapper[4828]: I1210 19:25:01.716291 4828 generic.go:334] "Generic (PLEG): container finished" podID="ca9f07ac-e1bc-42d2-9f45-c3b4b3843097" containerID="73625e2e7afed03b8010a8c1eca56d7ffb106f26c1523e5fa675968974984262" exitCode=0 Dec 10 19:25:01 crc kubenswrapper[4828]: I1210 19:25:01.716521 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2hgw" event={"ID":"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097","Type":"ContainerDied","Data":"73625e2e7afed03b8010a8c1eca56d7ffb106f26c1523e5fa675968974984262"} Dec 10 19:25:01 crc kubenswrapper[4828]: I1210 19:25:01.720263 4828 generic.go:334] "Generic (PLEG): container finished" podID="74792a2b-341a-4065-93e3-215c325db10c" containerID="db3ac48041235c638079fa2fe5577eb248918091b89d8da15c3b99b484b075b9" exitCode=0 Dec 10 19:25:01 crc kubenswrapper[4828]: I1210 19:25:01.720300 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" event={"ID":"74792a2b-341a-4065-93e3-215c325db10c","Type":"ContainerDied","Data":"db3ac48041235c638079fa2fe5577eb248918091b89d8da15c3b99b484b075b9"} Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.024614 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.145783 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-config\") pod \"74792a2b-341a-4065-93e3-215c325db10c\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.145973 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-dns-swift-storage-0\") pod \"74792a2b-341a-4065-93e3-215c325db10c\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.146002 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlf6b\" (UniqueName: \"kubernetes.io/projected/74792a2b-341a-4065-93e3-215c325db10c-kube-api-access-zlf6b\") pod \"74792a2b-341a-4065-93e3-215c325db10c\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.146723 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-ovsdbserver-nb\") pod \"74792a2b-341a-4065-93e3-215c325db10c\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.147135 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-openstack-edpm-ipam\") pod \"74792a2b-341a-4065-93e3-215c325db10c\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.147161 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-dns-svc\") pod \"74792a2b-341a-4065-93e3-215c325db10c\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.147223 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-ovsdbserver-sb\") pod \"74792a2b-341a-4065-93e3-215c325db10c\" (UID: \"74792a2b-341a-4065-93e3-215c325db10c\") " Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.152762 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74792a2b-341a-4065-93e3-215c325db10c-kube-api-access-zlf6b" (OuterVolumeSpecName: "kube-api-access-zlf6b") pod "74792a2b-341a-4065-93e3-215c325db10c" (UID: "74792a2b-341a-4065-93e3-215c325db10c"). InnerVolumeSpecName "kube-api-access-zlf6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.251196 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlf6b\" (UniqueName: \"kubernetes.io/projected/74792a2b-341a-4065-93e3-215c325db10c-kube-api-access-zlf6b\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.259231 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "74792a2b-341a-4065-93e3-215c325db10c" (UID: "74792a2b-341a-4065-93e3-215c325db10c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.260486 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "74792a2b-341a-4065-93e3-215c325db10c" (UID: "74792a2b-341a-4065-93e3-215c325db10c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.268383 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "74792a2b-341a-4065-93e3-215c325db10c" (UID: "74792a2b-341a-4065-93e3-215c325db10c"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.280219 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "74792a2b-341a-4065-93e3-215c325db10c" (UID: "74792a2b-341a-4065-93e3-215c325db10c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.315343 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "74792a2b-341a-4065-93e3-215c325db10c" (UID: "74792a2b-341a-4065-93e3-215c325db10c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.318877 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-config" (OuterVolumeSpecName: "config") pod "74792a2b-341a-4065-93e3-215c325db10c" (UID: "74792a2b-341a-4065-93e3-215c325db10c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.353839 4828 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-config\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.353907 4828 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.353923 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.353934 4828 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.353946 4828 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.353957 4828 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74792a2b-341a-4065-93e3-215c325db10c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.782887 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" event={"ID":"74792a2b-341a-4065-93e3-215c325db10c","Type":"ContainerDied","Data":"a6aa9472f68905cf0f279458f3a1f3de427c2533a06080e6dc8a5eeff5b23dde"} Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.782945 4828 scope.go:117] "RemoveContainer" containerID="db3ac48041235c638079fa2fe5577eb248918091b89d8da15c3b99b484b075b9" Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.783130 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-k6pwv" Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.848053 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-k6pwv"] Dec 10 19:25:03 crc kubenswrapper[4828]: I1210 19:25:03.860019 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-k6pwv"] Dec 10 19:25:04 crc kubenswrapper[4828]: I1210 19:25:04.425753 4828 scope.go:117] "RemoveContainer" containerID="8f5f1377dad6ef982d63e83d8849fe435180a62cf5b2650e456e971465502ae0" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.215794 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2"] Dec 10 19:25:05 crc kubenswrapper[4828]: E1210 19:25:05.216587 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74792a2b-341a-4065-93e3-215c325db10c" containerName="dnsmasq-dns" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.216603 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="74792a2b-341a-4065-93e3-215c325db10c" containerName="dnsmasq-dns" Dec 10 19:25:05 crc kubenswrapper[4828]: E1210 19:25:05.216695 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74792a2b-341a-4065-93e3-215c325db10c" containerName="init" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.216704 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="74792a2b-341a-4065-93e3-215c325db10c" containerName="init" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.216943 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="74792a2b-341a-4065-93e3-215c325db10c" containerName="dnsmasq-dns" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.217760 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.222094 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.222538 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.222676 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.223015 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.230694 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2"] Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.324920 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtkzg\" (UniqueName: \"kubernetes.io/projected/bd51b75d-403a-4b14-8f0f-72cd8d989bca-kube-api-access-wtkzg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2\" (UID: \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.325153 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2\" (UID: \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.325526 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2\" (UID: \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.325835 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2\" (UID: \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.428341 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2\" (UID: \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.428428 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtkzg\" (UniqueName: \"kubernetes.io/projected/bd51b75d-403a-4b14-8f0f-72cd8d989bca-kube-api-access-wtkzg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2\" (UID: \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.428510 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2\" (UID: \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.428642 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2\" (UID: \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.435792 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2\" (UID: \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.436905 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2\" (UID: \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.442196 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2\" (UID: \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.446334 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtkzg\" (UniqueName: \"kubernetes.io/projected/bd51b75d-403a-4b14-8f0f-72cd8d989bca-kube-api-access-wtkzg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2\" (UID: \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.543943 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.800576 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74792a2b-341a-4065-93e3-215c325db10c" path="/var/lib/kubelet/pods/74792a2b-341a-4065-93e3-215c325db10c/volumes" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.838419 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2hgw" event={"ID":"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097","Type":"ContainerStarted","Data":"1ea32bc24182edb6714104ebd4d9986fa194c5212ab8e684ed93d62f770b43af"} Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.844696 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cb5d96886-n5cp9" event={"ID":"d1887db4-171d-4bd1-8337-4a8c72e9e93d","Type":"ContainerStarted","Data":"fa0eed61414448332d509c5abec49baec5d263daf8baa4d5d505fc5fe44550dd"} Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.845824 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.848017 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5859bd5774-r9zl7" event={"ID":"961e506a-98a1-48e1-afd5-c4326af0bdb3","Type":"ContainerStarted","Data":"d9f42d1fc4e8535d4ce5310c5f0d6b9d711d56d6f1134a5d5d4804540e9cea2b"} Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.848551 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.859873 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j2hgw" podStartSLOduration=2.99815183 podStartE2EDuration="10.859856571s" podCreationTimestamp="2025-12-10 19:24:55 +0000 UTC" firstStartedPulling="2025-12-10 19:24:56.661716435 +0000 UTC m=+1777.172327440" lastFinishedPulling="2025-12-10 19:25:04.523421176 +0000 UTC m=+1785.034032181" observedRunningTime="2025-12-10 19:25:05.859132771 +0000 UTC m=+1786.369743776" watchObservedRunningTime="2025-12-10 19:25:05.859856571 +0000 UTC m=+1786.370467576" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.886275 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5859bd5774-r9zl7" podStartSLOduration=2.988114569 podStartE2EDuration="9.88625881s" podCreationTimestamp="2025-12-10 19:24:56 +0000 UTC" firstStartedPulling="2025-12-10 19:24:57.596377619 +0000 UTC m=+1778.106988624" lastFinishedPulling="2025-12-10 19:25:04.49452185 +0000 UTC m=+1785.005132865" observedRunningTime="2025-12-10 19:25:05.882706126 +0000 UTC m=+1786.393317131" watchObservedRunningTime="2025-12-10 19:25:05.88625881 +0000 UTC m=+1786.396869805" Dec 10 19:25:05 crc kubenswrapper[4828]: I1210 19:25:05.925022 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-cb5d96886-n5cp9" podStartSLOduration=3.216303631 podStartE2EDuration="9.925001416s" podCreationTimestamp="2025-12-10 19:24:56 +0000 UTC" firstStartedPulling="2025-12-10 19:24:57.789233566 +0000 UTC m=+1778.299844571" lastFinishedPulling="2025-12-10 19:25:04.497931351 +0000 UTC m=+1785.008542356" observedRunningTime="2025-12-10 19:25:05.902630213 +0000 UTC m=+1786.413241228" watchObservedRunningTime="2025-12-10 19:25:05.925001416 +0000 UTC m=+1786.435612421" Dec 10 19:25:06 crc kubenswrapper[4828]: I1210 19:25:06.217689 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2"] Dec 10 19:25:06 crc kubenswrapper[4828]: I1210 19:25:06.864290 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" event={"ID":"bd51b75d-403a-4b14-8f0f-72cd8d989bca","Type":"ContainerStarted","Data":"600503f7ce8f6d7620ea5b5f3cb0bad892f750d4dacd29165a784fad4e9aa44f"} Dec 10 19:25:07 crc kubenswrapper[4828]: I1210 19:25:07.794453 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:25:07 crc kubenswrapper[4828]: E1210 19:25:07.795362 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:25:13 crc kubenswrapper[4828]: I1210 19:25:13.992739 4828 generic.go:334] "Generic (PLEG): container finished" podID="3eb66f68-0443-4122-8d37-8000aefcbc3b" containerID="81b437b4051791af4f516370cff7655f490ad68da013321b50f0b02df1c4abc1" exitCode=0 Dec 10 19:25:13 crc kubenswrapper[4828]: I1210 19:25:13.992939 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3eb66f68-0443-4122-8d37-8000aefcbc3b","Type":"ContainerDied","Data":"81b437b4051791af4f516370cff7655f490ad68da013321b50f0b02df1c4abc1"} Dec 10 19:25:14 crc kubenswrapper[4828]: I1210 19:25:14.025980 4828 generic.go:334] "Generic (PLEG): container finished" podID="540f0868-0b12-44cc-806a-9d31d644da25" containerID="bb6e22e205bb8ee4d62d7b1021e1ea28e7d87c0650c394c1fa096feafbbd5683" exitCode=0 Dec 10 19:25:14 crc kubenswrapper[4828]: I1210 19:25:14.026027 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"540f0868-0b12-44cc-806a-9d31d644da25","Type":"ContainerDied","Data":"bb6e22e205bb8ee4d62d7b1021e1ea28e7d87c0650c394c1fa096feafbbd5683"} Dec 10 19:25:15 crc kubenswrapper[4828]: I1210 19:25:15.488780 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-5859bd5774-r9zl7" Dec 10 19:25:15 crc kubenswrapper[4828]: I1210 19:25:15.502862 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j2hgw" Dec 10 19:25:15 crc kubenswrapper[4828]: I1210 19:25:15.502997 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j2hgw" Dec 10 19:25:15 crc kubenswrapper[4828]: I1210 19:25:15.523563 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-cb5d96886-n5cp9" Dec 10 19:25:15 crc kubenswrapper[4828]: I1210 19:25:15.592842 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-59c4ffddfc-nf4zg"] Dec 10 19:25:15 crc kubenswrapper[4828]: I1210 19:25:15.593071 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-59c4ffddfc-nf4zg" podUID="866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709" containerName="heat-api" containerID="cri-o://4e62436d5f048b277525938d4b6e74b8b4d6af7192d374e82a9fb90de498c848" gracePeriod=60 Dec 10 19:25:15 crc kubenswrapper[4828]: I1210 19:25:15.593885 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j2hgw" Dec 10 19:25:15 crc kubenswrapper[4828]: I1210 19:25:15.608354 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-546dfdfc9f-s9ctn"] Dec 10 19:25:15 crc kubenswrapper[4828]: I1210 19:25:15.608566 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" podUID="296faf68-6251-48d1-bc36-026a31c00393" containerName="heat-cfnapi" containerID="cri-o://b514a4dbaafec84cc689eb9deaf63a5685ff0cb779acc7f815431dae3f169013" gracePeriod=60 Dec 10 19:25:16 crc kubenswrapper[4828]: I1210 19:25:16.123767 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j2hgw" Dec 10 19:25:16 crc kubenswrapper[4828]: I1210 19:25:16.173117 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j2hgw"] Dec 10 19:25:16 crc kubenswrapper[4828]: I1210 19:25:16.972031 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-55d5c67ddc-vcr2w" Dec 10 19:25:17 crc kubenswrapper[4828]: I1210 19:25:17.028397 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-d78dbcd66-dq2t9"] Dec 10 19:25:17 crc kubenswrapper[4828]: I1210 19:25:17.028614 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-d78dbcd66-dq2t9" podUID="7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b" containerName="heat-engine" containerID="cri-o://3f7f00335c86a1e5b99e0e3677a9c6cf168454f042fc2f5a720193b77fff2c7e" gracePeriod=60 Dec 10 19:25:17 crc kubenswrapper[4828]: I1210 19:25:17.069233 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" event={"ID":"bd51b75d-403a-4b14-8f0f-72cd8d989bca","Type":"ContainerStarted","Data":"c125b87ce42327f0f1f720fbf6dbcee1f0ee75d7d9d77aaed097cac5165e23e1"} Dec 10 19:25:17 crc kubenswrapper[4828]: I1210 19:25:17.071885 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3eb66f68-0443-4122-8d37-8000aefcbc3b","Type":"ContainerStarted","Data":"1fef5806c76d0785e6c3cdb4110e573a409ec0131d1b6722da9a9f1e43997199"} Dec 10 19:25:17 crc kubenswrapper[4828]: I1210 19:25:17.072101 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:25:17 crc kubenswrapper[4828]: I1210 19:25:17.074426 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"540f0868-0b12-44cc-806a-9d31d644da25","Type":"ContainerStarted","Data":"1c4450d61918b597a1ba7c8058d6e724f9dde14beeec5e6bc399888f594836b9"} Dec 10 19:25:17 crc kubenswrapper[4828]: I1210 19:25:17.074713 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 10 19:25:17 crc kubenswrapper[4828]: I1210 19:25:17.092568 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" podStartSLOduration=2.5592794359999997 podStartE2EDuration="12.092546217s" podCreationTimestamp="2025-12-10 19:25:05 +0000 UTC" firstStartedPulling="2025-12-10 19:25:06.221140748 +0000 UTC m=+1786.731751753" lastFinishedPulling="2025-12-10 19:25:15.754407529 +0000 UTC m=+1796.265018534" observedRunningTime="2025-12-10 19:25:17.085352557 +0000 UTC m=+1797.595963572" watchObservedRunningTime="2025-12-10 19:25:17.092546217 +0000 UTC m=+1797.603157222" Dec 10 19:25:17 crc kubenswrapper[4828]: I1210 19:25:17.119639 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=41.119613144 podStartE2EDuration="41.119613144s" podCreationTimestamp="2025-12-10 19:24:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:25:17.109571818 +0000 UTC m=+1797.620182823" watchObservedRunningTime="2025-12-10 19:25:17.119613144 +0000 UTC m=+1797.630224149" Dec 10 19:25:17 crc kubenswrapper[4828]: I1210 19:25:17.146748 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=41.146725742 podStartE2EDuration="41.146725742s" podCreationTimestamp="2025-12-10 19:24:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 19:25:17.143059115 +0000 UTC m=+1797.653670120" watchObservedRunningTime="2025-12-10 19:25:17.146725742 +0000 UTC m=+1797.657336747" Dec 10 19:25:18 crc kubenswrapper[4828]: I1210 19:25:18.083735 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j2hgw" podUID="ca9f07ac-e1bc-42d2-9f45-c3b4b3843097" containerName="registry-server" containerID="cri-o://1ea32bc24182edb6714104ebd4d9986fa194c5212ab8e684ed93d62f770b43af" gracePeriod=2 Dec 10 19:25:18 crc kubenswrapper[4828]: I1210 19:25:18.602831 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j2hgw" Dec 10 19:25:18 crc kubenswrapper[4828]: I1210 19:25:18.721831 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fzt6\" (UniqueName: \"kubernetes.io/projected/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-kube-api-access-8fzt6\") pod \"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097\" (UID: \"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097\") " Dec 10 19:25:18 crc kubenswrapper[4828]: I1210 19:25:18.721955 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-catalog-content\") pod \"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097\" (UID: \"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097\") " Dec 10 19:25:18 crc kubenswrapper[4828]: I1210 19:25:18.722144 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-utilities\") pod \"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097\" (UID: \"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097\") " Dec 10 19:25:18 crc kubenswrapper[4828]: I1210 19:25:18.722995 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-utilities" (OuterVolumeSpecName: "utilities") pod "ca9f07ac-e1bc-42d2-9f45-c3b4b3843097" (UID: "ca9f07ac-e1bc-42d2-9f45-c3b4b3843097"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:25:18 crc kubenswrapper[4828]: I1210 19:25:18.723258 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:18 crc kubenswrapper[4828]: I1210 19:25:18.728435 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-kube-api-access-8fzt6" (OuterVolumeSpecName: "kube-api-access-8fzt6") pod "ca9f07ac-e1bc-42d2-9f45-c3b4b3843097" (UID: "ca9f07ac-e1bc-42d2-9f45-c3b4b3843097"). InnerVolumeSpecName "kube-api-access-8fzt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:25:18 crc kubenswrapper[4828]: I1210 19:25:18.782878 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" podUID="296faf68-6251-48d1-bc36-026a31c00393" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.0.221:8000/healthcheck\": read tcp 10.217.0.2:58614->10.217.0.221:8000: read: connection reset by peer" Dec 10 19:25:18 crc kubenswrapper[4828]: I1210 19:25:18.790389 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca9f07ac-e1bc-42d2-9f45-c3b4b3843097" (UID: "ca9f07ac-e1bc-42d2-9f45-c3b4b3843097"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:25:18 crc kubenswrapper[4828]: I1210 19:25:18.825741 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fzt6\" (UniqueName: \"kubernetes.io/projected/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-kube-api-access-8fzt6\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:18 crc kubenswrapper[4828]: I1210 19:25:18.825779 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.071265 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-59c4ffddfc-nf4zg" podUID="866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.0.220:8004/healthcheck\": read tcp 10.217.0.2:33768->10.217.0.220:8004: read: connection reset by peer" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.149061 4828 generic.go:334] "Generic (PLEG): container finished" podID="296faf68-6251-48d1-bc36-026a31c00393" containerID="b514a4dbaafec84cc689eb9deaf63a5685ff0cb779acc7f815431dae3f169013" exitCode=0 Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.149767 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" event={"ID":"296faf68-6251-48d1-bc36-026a31c00393","Type":"ContainerDied","Data":"b514a4dbaafec84cc689eb9deaf63a5685ff0cb779acc7f815431dae3f169013"} Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.172099 4828 generic.go:334] "Generic (PLEG): container finished" podID="ca9f07ac-e1bc-42d2-9f45-c3b4b3843097" containerID="1ea32bc24182edb6714104ebd4d9986fa194c5212ab8e684ed93d62f770b43af" exitCode=0 Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.172168 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2hgw" event={"ID":"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097","Type":"ContainerDied","Data":"1ea32bc24182edb6714104ebd4d9986fa194c5212ab8e684ed93d62f770b43af"} Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.172202 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j2hgw" event={"ID":"ca9f07ac-e1bc-42d2-9f45-c3b4b3843097","Type":"ContainerDied","Data":"342e5640c0465e251d7c8245913f75bf07c1c6eebe6d71a4bd5c7c899e0f4e93"} Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.172220 4828 scope.go:117] "RemoveContainer" containerID="1ea32bc24182edb6714104ebd4d9986fa194c5212ab8e684ed93d62f770b43af" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.172456 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j2hgw" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.234003 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j2hgw"] Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.234199 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.245713 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j2hgw"] Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.247930 4828 scope.go:117] "RemoveContainer" containerID="73625e2e7afed03b8010a8c1eca56d7ffb106f26c1523e5fa675968974984262" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.331316 4828 scope.go:117] "RemoveContainer" containerID="0056fff6b7bc9d122721ca99f1581024ab0338687fec8f4aa7dd1618525d7019" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.343908 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-config-data-custom\") pod \"296faf68-6251-48d1-bc36-026a31c00393\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.343969 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-combined-ca-bundle\") pod \"296faf68-6251-48d1-bc36-026a31c00393\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.343990 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-internal-tls-certs\") pod \"296faf68-6251-48d1-bc36-026a31c00393\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.344026 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-config-data\") pod \"296faf68-6251-48d1-bc36-026a31c00393\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.344123 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-public-tls-certs\") pod \"296faf68-6251-48d1-bc36-026a31c00393\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.344174 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8v5r\" (UniqueName: \"kubernetes.io/projected/296faf68-6251-48d1-bc36-026a31c00393-kube-api-access-t8v5r\") pod \"296faf68-6251-48d1-bc36-026a31c00393\" (UID: \"296faf68-6251-48d1-bc36-026a31c00393\") " Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.378574 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/296faf68-6251-48d1-bc36-026a31c00393-kube-api-access-t8v5r" (OuterVolumeSpecName: "kube-api-access-t8v5r") pod "296faf68-6251-48d1-bc36-026a31c00393" (UID: "296faf68-6251-48d1-bc36-026a31c00393"). InnerVolumeSpecName "kube-api-access-t8v5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.395015 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "296faf68-6251-48d1-bc36-026a31c00393" (UID: "296faf68-6251-48d1-bc36-026a31c00393"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.447094 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8v5r\" (UniqueName: \"kubernetes.io/projected/296faf68-6251-48d1-bc36-026a31c00393-kube-api-access-t8v5r\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.447136 4828 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.474035 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "296faf68-6251-48d1-bc36-026a31c00393" (UID: "296faf68-6251-48d1-bc36-026a31c00393"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.497283 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "296faf68-6251-48d1-bc36-026a31c00393" (UID: "296faf68-6251-48d1-bc36-026a31c00393"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.545727 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-config-data" (OuterVolumeSpecName: "config-data") pod "296faf68-6251-48d1-bc36-026a31c00393" (UID: "296faf68-6251-48d1-bc36-026a31c00393"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.549254 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.549370 4828 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.549444 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.550781 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "296faf68-6251-48d1-bc36-026a31c00393" (UID: "296faf68-6251-48d1-bc36-026a31c00393"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.557540 4828 scope.go:117] "RemoveContainer" containerID="1ea32bc24182edb6714104ebd4d9986fa194c5212ab8e684ed93d62f770b43af" Dec 10 19:25:19 crc kubenswrapper[4828]: E1210 19:25:19.561575 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ea32bc24182edb6714104ebd4d9986fa194c5212ab8e684ed93d62f770b43af\": container with ID starting with 1ea32bc24182edb6714104ebd4d9986fa194c5212ab8e684ed93d62f770b43af not found: ID does not exist" containerID="1ea32bc24182edb6714104ebd4d9986fa194c5212ab8e684ed93d62f770b43af" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.561624 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ea32bc24182edb6714104ebd4d9986fa194c5212ab8e684ed93d62f770b43af"} err="failed to get container status \"1ea32bc24182edb6714104ebd4d9986fa194c5212ab8e684ed93d62f770b43af\": rpc error: code = NotFound desc = could not find container \"1ea32bc24182edb6714104ebd4d9986fa194c5212ab8e684ed93d62f770b43af\": container with ID starting with 1ea32bc24182edb6714104ebd4d9986fa194c5212ab8e684ed93d62f770b43af not found: ID does not exist" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.561648 4828 scope.go:117] "RemoveContainer" containerID="73625e2e7afed03b8010a8c1eca56d7ffb106f26c1523e5fa675968974984262" Dec 10 19:25:19 crc kubenswrapper[4828]: E1210 19:25:19.562050 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73625e2e7afed03b8010a8c1eca56d7ffb106f26c1523e5fa675968974984262\": container with ID starting with 73625e2e7afed03b8010a8c1eca56d7ffb106f26c1523e5fa675968974984262 not found: ID does not exist" containerID="73625e2e7afed03b8010a8c1eca56d7ffb106f26c1523e5fa675968974984262" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.562074 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73625e2e7afed03b8010a8c1eca56d7ffb106f26c1523e5fa675968974984262"} err="failed to get container status \"73625e2e7afed03b8010a8c1eca56d7ffb106f26c1523e5fa675968974984262\": rpc error: code = NotFound desc = could not find container \"73625e2e7afed03b8010a8c1eca56d7ffb106f26c1523e5fa675968974984262\": container with ID starting with 73625e2e7afed03b8010a8c1eca56d7ffb106f26c1523e5fa675968974984262 not found: ID does not exist" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.562088 4828 scope.go:117] "RemoveContainer" containerID="0056fff6b7bc9d122721ca99f1581024ab0338687fec8f4aa7dd1618525d7019" Dec 10 19:25:19 crc kubenswrapper[4828]: E1210 19:25:19.562600 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0056fff6b7bc9d122721ca99f1581024ab0338687fec8f4aa7dd1618525d7019\": container with ID starting with 0056fff6b7bc9d122721ca99f1581024ab0338687fec8f4aa7dd1618525d7019 not found: ID does not exist" containerID="0056fff6b7bc9d122721ca99f1581024ab0338687fec8f4aa7dd1618525d7019" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.562625 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0056fff6b7bc9d122721ca99f1581024ab0338687fec8f4aa7dd1618525d7019"} err="failed to get container status \"0056fff6b7bc9d122721ca99f1581024ab0338687fec8f4aa7dd1618525d7019\": rpc error: code = NotFound desc = could not find container \"0056fff6b7bc9d122721ca99f1581024ab0338687fec8f4aa7dd1618525d7019\": container with ID starting with 0056fff6b7bc9d122721ca99f1581024ab0338687fec8f4aa7dd1618525d7019 not found: ID does not exist" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.652913 4828 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/296faf68-6251-48d1-bc36-026a31c00393-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.699094 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.754580 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-combined-ca-bundle\") pod \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.754721 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-config-data\") pod \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.754787 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-internal-tls-certs\") pod \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.754836 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lptvz\" (UniqueName: \"kubernetes.io/projected/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-kube-api-access-lptvz\") pod \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.754871 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-public-tls-certs\") pod \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.754911 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-config-data-custom\") pod \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\" (UID: \"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709\") " Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.762100 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709" (UID: "866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.764939 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-kube-api-access-lptvz" (OuterVolumeSpecName: "kube-api-access-lptvz") pod "866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709" (UID: "866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709"). InnerVolumeSpecName "kube-api-access-lptvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.810383 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca9f07ac-e1bc-42d2-9f45-c3b4b3843097" path="/var/lib/kubelet/pods/ca9f07ac-e1bc-42d2-9f45-c3b4b3843097/volumes" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.837977 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709" (UID: "866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.857763 4828 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.857849 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.857859 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lptvz\" (UniqueName: \"kubernetes.io/projected/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-kube-api-access-lptvz\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.864020 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709" (UID: "866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.864204 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-config-data" (OuterVolumeSpecName: "config-data") pod "866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709" (UID: "866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.867420 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709" (UID: "866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.960158 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.960194 4828 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:19 crc kubenswrapper[4828]: I1210 19:25:19.960209 4828 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:20 crc kubenswrapper[4828]: I1210 19:25:20.185255 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" event={"ID":"296faf68-6251-48d1-bc36-026a31c00393","Type":"ContainerDied","Data":"0bb5e90efc664c51989f69ee2d7abf7b0061f2d85cb18eaf58cfdd7c0d1065bf"} Dec 10 19:25:20 crc kubenswrapper[4828]: I1210 19:25:20.185645 4828 scope.go:117] "RemoveContainer" containerID="b514a4dbaafec84cc689eb9deaf63a5685ff0cb779acc7f815431dae3f169013" Dec 10 19:25:20 crc kubenswrapper[4828]: I1210 19:25:20.185273 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-546dfdfc9f-s9ctn" Dec 10 19:25:20 crc kubenswrapper[4828]: I1210 19:25:20.189141 4828 generic.go:334] "Generic (PLEG): container finished" podID="866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709" containerID="4e62436d5f048b277525938d4b6e74b8b4d6af7192d374e82a9fb90de498c848" exitCode=0 Dec 10 19:25:20 crc kubenswrapper[4828]: I1210 19:25:20.189212 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-59c4ffddfc-nf4zg" event={"ID":"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709","Type":"ContainerDied","Data":"4e62436d5f048b277525938d4b6e74b8b4d6af7192d374e82a9fb90de498c848"} Dec 10 19:25:20 crc kubenswrapper[4828]: I1210 19:25:20.189245 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-59c4ffddfc-nf4zg" event={"ID":"866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709","Type":"ContainerDied","Data":"9449e358d2c50c03042939037ad52b848ccaf0d226e7ddd9f232bb8cc911cb58"} Dec 10 19:25:20 crc kubenswrapper[4828]: I1210 19:25:20.189311 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-59c4ffddfc-nf4zg" Dec 10 19:25:20 crc kubenswrapper[4828]: I1210 19:25:20.233012 4828 scope.go:117] "RemoveContainer" containerID="4e62436d5f048b277525938d4b6e74b8b4d6af7192d374e82a9fb90de498c848" Dec 10 19:25:20 crc kubenswrapper[4828]: I1210 19:25:20.249145 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-546dfdfc9f-s9ctn"] Dec 10 19:25:20 crc kubenswrapper[4828]: I1210 19:25:20.262336 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-546dfdfc9f-s9ctn"] Dec 10 19:25:20 crc kubenswrapper[4828]: I1210 19:25:20.274337 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-59c4ffddfc-nf4zg"] Dec 10 19:25:20 crc kubenswrapper[4828]: I1210 19:25:20.288844 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-59c4ffddfc-nf4zg"] Dec 10 19:25:20 crc kubenswrapper[4828]: I1210 19:25:20.387103 4828 scope.go:117] "RemoveContainer" containerID="4e62436d5f048b277525938d4b6e74b8b4d6af7192d374e82a9fb90de498c848" Dec 10 19:25:20 crc kubenswrapper[4828]: E1210 19:25:20.387607 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e62436d5f048b277525938d4b6e74b8b4d6af7192d374e82a9fb90de498c848\": container with ID starting with 4e62436d5f048b277525938d4b6e74b8b4d6af7192d374e82a9fb90de498c848 not found: ID does not exist" containerID="4e62436d5f048b277525938d4b6e74b8b4d6af7192d374e82a9fb90de498c848" Dec 10 19:25:20 crc kubenswrapper[4828]: I1210 19:25:20.387639 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e62436d5f048b277525938d4b6e74b8b4d6af7192d374e82a9fb90de498c848"} err="failed to get container status \"4e62436d5f048b277525938d4b6e74b8b4d6af7192d374e82a9fb90de498c848\": rpc error: code = NotFound desc = could not find container \"4e62436d5f048b277525938d4b6e74b8b4d6af7192d374e82a9fb90de498c848\": container with ID starting with 4e62436d5f048b277525938d4b6e74b8b4d6af7192d374e82a9fb90de498c848 not found: ID does not exist" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.313770 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-dtpxc"] Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.332471 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-dtpxc"] Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.418411 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-7krk4"] Dec 10 19:25:21 crc kubenswrapper[4828]: E1210 19:25:21.419094 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709" containerName="heat-api" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.419115 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709" containerName="heat-api" Dec 10 19:25:21 crc kubenswrapper[4828]: E1210 19:25:21.419138 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca9f07ac-e1bc-42d2-9f45-c3b4b3843097" containerName="extract-content" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.419146 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca9f07ac-e1bc-42d2-9f45-c3b4b3843097" containerName="extract-content" Dec 10 19:25:21 crc kubenswrapper[4828]: E1210 19:25:21.419167 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca9f07ac-e1bc-42d2-9f45-c3b4b3843097" containerName="extract-utilities" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.419177 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca9f07ac-e1bc-42d2-9f45-c3b4b3843097" containerName="extract-utilities" Dec 10 19:25:21 crc kubenswrapper[4828]: E1210 19:25:21.419200 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="296faf68-6251-48d1-bc36-026a31c00393" containerName="heat-cfnapi" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.419208 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="296faf68-6251-48d1-bc36-026a31c00393" containerName="heat-cfnapi" Dec 10 19:25:21 crc kubenswrapper[4828]: E1210 19:25:21.419231 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca9f07ac-e1bc-42d2-9f45-c3b4b3843097" containerName="registry-server" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.419242 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca9f07ac-e1bc-42d2-9f45-c3b4b3843097" containerName="registry-server" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.419512 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="296faf68-6251-48d1-bc36-026a31c00393" containerName="heat-cfnapi" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.419535 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709" containerName="heat-api" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.419555 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca9f07ac-e1bc-42d2-9f45-c3b4b3843097" containerName="registry-server" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.420664 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-7krk4" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.422456 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.441417 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-7krk4"] Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.495409 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-scripts\") pod \"aodh-db-sync-7krk4\" (UID: \"788b1631-a897-4929-972f-4c8b9bf008db\") " pod="openstack/aodh-db-sync-7krk4" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.495514 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-config-data\") pod \"aodh-db-sync-7krk4\" (UID: \"788b1631-a897-4929-972f-4c8b9bf008db\") " pod="openstack/aodh-db-sync-7krk4" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.495537 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-combined-ca-bundle\") pod \"aodh-db-sync-7krk4\" (UID: \"788b1631-a897-4929-972f-4c8b9bf008db\") " pod="openstack/aodh-db-sync-7krk4" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.495983 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjn9x\" (UniqueName: \"kubernetes.io/projected/788b1631-a897-4929-972f-4c8b9bf008db-kube-api-access-kjn9x\") pod \"aodh-db-sync-7krk4\" (UID: \"788b1631-a897-4929-972f-4c8b9bf008db\") " pod="openstack/aodh-db-sync-7krk4" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.599124 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjn9x\" (UniqueName: \"kubernetes.io/projected/788b1631-a897-4929-972f-4c8b9bf008db-kube-api-access-kjn9x\") pod \"aodh-db-sync-7krk4\" (UID: \"788b1631-a897-4929-972f-4c8b9bf008db\") " pod="openstack/aodh-db-sync-7krk4" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.599338 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-scripts\") pod \"aodh-db-sync-7krk4\" (UID: \"788b1631-a897-4929-972f-4c8b9bf008db\") " pod="openstack/aodh-db-sync-7krk4" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.599406 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-config-data\") pod \"aodh-db-sync-7krk4\" (UID: \"788b1631-a897-4929-972f-4c8b9bf008db\") " pod="openstack/aodh-db-sync-7krk4" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.599435 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-combined-ca-bundle\") pod \"aodh-db-sync-7krk4\" (UID: \"788b1631-a897-4929-972f-4c8b9bf008db\") " pod="openstack/aodh-db-sync-7krk4" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.604135 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-scripts\") pod \"aodh-db-sync-7krk4\" (UID: \"788b1631-a897-4929-972f-4c8b9bf008db\") " pod="openstack/aodh-db-sync-7krk4" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.606103 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-combined-ca-bundle\") pod \"aodh-db-sync-7krk4\" (UID: \"788b1631-a897-4929-972f-4c8b9bf008db\") " pod="openstack/aodh-db-sync-7krk4" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.606736 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-config-data\") pod \"aodh-db-sync-7krk4\" (UID: \"788b1631-a897-4929-972f-4c8b9bf008db\") " pod="openstack/aodh-db-sync-7krk4" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.622160 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjn9x\" (UniqueName: \"kubernetes.io/projected/788b1631-a897-4929-972f-4c8b9bf008db-kube-api-access-kjn9x\") pod \"aodh-db-sync-7krk4\" (UID: \"788b1631-a897-4929-972f-4c8b9bf008db\") " pod="openstack/aodh-db-sync-7krk4" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.746009 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-7krk4" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.789393 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:25:21 crc kubenswrapper[4828]: E1210 19:25:21.790003 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.819091 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="057a6200-117a-47be-97df-55ae94e6f763" path="/var/lib/kubelet/pods/057a6200-117a-47be-97df-55ae94e6f763/volumes" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.819883 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="296faf68-6251-48d1-bc36-026a31c00393" path="/var/lib/kubelet/pods/296faf68-6251-48d1-bc36-026a31c00393/volumes" Dec 10 19:25:21 crc kubenswrapper[4828]: I1210 19:25:21.820595 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709" path="/var/lib/kubelet/pods/866f8d3c-5bc0-4b03-b20c-f7ce4aeb0709/volumes" Dec 10 19:25:22 crc kubenswrapper[4828]: W1210 19:25:22.237816 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod788b1631_a897_4929_972f_4c8b9bf008db.slice/crio-d7b21c0329861f4f2b3afbeb34739951ae9ef6a4d3db6b450945c4917ebccc35 WatchSource:0}: Error finding container d7b21c0329861f4f2b3afbeb34739951ae9ef6a4d3db6b450945c4917ebccc35: Status 404 returned error can't find the container with id d7b21c0329861f4f2b3afbeb34739951ae9ef6a4d3db6b450945c4917ebccc35 Dec 10 19:25:22 crc kubenswrapper[4828]: I1210 19:25:22.238094 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-7krk4"] Dec 10 19:25:23 crc kubenswrapper[4828]: I1210 19:25:23.264607 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-7krk4" event={"ID":"788b1631-a897-4929-972f-4c8b9bf008db","Type":"ContainerStarted","Data":"d7b21c0329861f4f2b3afbeb34739951ae9ef6a4d3db6b450945c4917ebccc35"} Dec 10 19:25:23 crc kubenswrapper[4828]: E1210 19:25:23.699383 4828 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3f7f00335c86a1e5b99e0e3677a9c6cf168454f042fc2f5a720193b77fff2c7e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 10 19:25:23 crc kubenswrapper[4828]: E1210 19:25:23.701377 4828 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3f7f00335c86a1e5b99e0e3677a9c6cf168454f042fc2f5a720193b77fff2c7e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 10 19:25:23 crc kubenswrapper[4828]: E1210 19:25:23.703710 4828 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3f7f00335c86a1e5b99e0e3677a9c6cf168454f042fc2f5a720193b77fff2c7e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 10 19:25:23 crc kubenswrapper[4828]: E1210 19:25:23.703763 4828 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-d78dbcd66-dq2t9" podUID="7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b" containerName="heat-engine" Dec 10 19:25:27 crc kubenswrapper[4828]: I1210 19:25:27.340278 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-7krk4" event={"ID":"788b1631-a897-4929-972f-4c8b9bf008db","Type":"ContainerStarted","Data":"83c59f81f6ea289aaabe02c0e97b3bb1998fcf5e568319c373edf91d5ae2c18b"} Dec 10 19:25:27 crc kubenswrapper[4828]: I1210 19:25:27.363549 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-7krk4" podStartSLOduration=2.505709726 podStartE2EDuration="6.363531567s" podCreationTimestamp="2025-12-10 19:25:21 +0000 UTC" firstStartedPulling="2025-12-10 19:25:22.245283614 +0000 UTC m=+1802.755894609" lastFinishedPulling="2025-12-10 19:25:26.103105425 +0000 UTC m=+1806.613716450" observedRunningTime="2025-12-10 19:25:27.354316672 +0000 UTC m=+1807.864927677" watchObservedRunningTime="2025-12-10 19:25:27.363531567 +0000 UTC m=+1807.874142562" Dec 10 19:25:27 crc kubenswrapper[4828]: I1210 19:25:27.501350 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 10 19:25:27 crc kubenswrapper[4828]: I1210 19:25:27.586375 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.367844 4828 generic.go:334] "Generic (PLEG): container finished" podID="7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b" containerID="3f7f00335c86a1e5b99e0e3677a9c6cf168454f042fc2f5a720193b77fff2c7e" exitCode=0 Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.368571 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-d78dbcd66-dq2t9" event={"ID":"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b","Type":"ContainerDied","Data":"3f7f00335c86a1e5b99e0e3677a9c6cf168454f042fc2f5a720193b77fff2c7e"} Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.368603 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-d78dbcd66-dq2t9" event={"ID":"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b","Type":"ContainerDied","Data":"b962bcb63927ba9ce27b85c9abfe3ecabe99b6d3c11f40facbbaeea501982b41"} Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.368613 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b962bcb63927ba9ce27b85c9abfe3ecabe99b6d3c11f40facbbaeea501982b41" Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.371644 4828 generic.go:334] "Generic (PLEG): container finished" podID="788b1631-a897-4929-972f-4c8b9bf008db" containerID="83c59f81f6ea289aaabe02c0e97b3bb1998fcf5e568319c373edf91d5ae2c18b" exitCode=0 Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.371757 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-7krk4" event={"ID":"788b1631-a897-4929-972f-4c8b9bf008db","Type":"ContainerDied","Data":"83c59f81f6ea289aaabe02c0e97b3bb1998fcf5e568319c373edf91d5ae2c18b"} Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.373838 4828 generic.go:334] "Generic (PLEG): container finished" podID="bd51b75d-403a-4b14-8f0f-72cd8d989bca" containerID="c125b87ce42327f0f1f720fbf6dbcee1f0ee75d7d9d77aaed097cac5165e23e1" exitCode=0 Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.373970 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" event={"ID":"bd51b75d-403a-4b14-8f0f-72cd8d989bca","Type":"ContainerDied","Data":"c125b87ce42327f0f1f720fbf6dbcee1f0ee75d7d9d77aaed097cac5165e23e1"} Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.457418 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.509971 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-config-data-custom\") pod \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\" (UID: \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\") " Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.510017 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-config-data\") pod \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\" (UID: \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\") " Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.510051 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc9kc\" (UniqueName: \"kubernetes.io/projected/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-kube-api-access-fc9kc\") pod \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\" (UID: \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\") " Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.510119 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-combined-ca-bundle\") pod \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\" (UID: \"7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b\") " Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.520372 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b" (UID: "7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.525051 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-kube-api-access-fc9kc" (OuterVolumeSpecName: "kube-api-access-fc9kc") pod "7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b" (UID: "7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b"). InnerVolumeSpecName "kube-api-access-fc9kc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.550127 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b" (UID: "7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.595038 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-config-data" (OuterVolumeSpecName: "config-data") pod "7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b" (UID: "7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.613272 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc9kc\" (UniqueName: \"kubernetes.io/projected/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-kube-api-access-fc9kc\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.613612 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.613687 4828 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:29 crc kubenswrapper[4828]: I1210 19:25:29.613773 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:30 crc kubenswrapper[4828]: I1210 19:25:30.384385 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-d78dbcd66-dq2t9" Dec 10 19:25:30 crc kubenswrapper[4828]: I1210 19:25:30.419241 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-d78dbcd66-dq2t9"] Dec 10 19:25:30 crc kubenswrapper[4828]: I1210 19:25:30.437854 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-d78dbcd66-dq2t9"] Dec 10 19:25:30 crc kubenswrapper[4828]: I1210 19:25:30.807676 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-7krk4" Dec 10 19:25:30 crc kubenswrapper[4828]: I1210 19:25:30.915062 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" Dec 10 19:25:30 crc kubenswrapper[4828]: I1210 19:25:30.943397 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-combined-ca-bundle\") pod \"788b1631-a897-4929-972f-4c8b9bf008db\" (UID: \"788b1631-a897-4929-972f-4c8b9bf008db\") " Dec 10 19:25:30 crc kubenswrapper[4828]: I1210 19:25:30.943509 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjn9x\" (UniqueName: \"kubernetes.io/projected/788b1631-a897-4929-972f-4c8b9bf008db-kube-api-access-kjn9x\") pod \"788b1631-a897-4929-972f-4c8b9bf008db\" (UID: \"788b1631-a897-4929-972f-4c8b9bf008db\") " Dec 10 19:25:30 crc kubenswrapper[4828]: I1210 19:25:30.943572 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-config-data\") pod \"788b1631-a897-4929-972f-4c8b9bf008db\" (UID: \"788b1631-a897-4929-972f-4c8b9bf008db\") " Dec 10 19:25:30 crc kubenswrapper[4828]: I1210 19:25:30.943639 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-scripts\") pod \"788b1631-a897-4929-972f-4c8b9bf008db\" (UID: \"788b1631-a897-4929-972f-4c8b9bf008db\") " Dec 10 19:25:30 crc kubenswrapper[4828]: I1210 19:25:30.948671 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-scripts" (OuterVolumeSpecName: "scripts") pod "788b1631-a897-4929-972f-4c8b9bf008db" (UID: "788b1631-a897-4929-972f-4c8b9bf008db"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:30 crc kubenswrapper[4828]: I1210 19:25:30.948901 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/788b1631-a897-4929-972f-4c8b9bf008db-kube-api-access-kjn9x" (OuterVolumeSpecName: "kube-api-access-kjn9x") pod "788b1631-a897-4929-972f-4c8b9bf008db" (UID: "788b1631-a897-4929-972f-4c8b9bf008db"). InnerVolumeSpecName "kube-api-access-kjn9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:25:30 crc kubenswrapper[4828]: I1210 19:25:30.978703 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "788b1631-a897-4929-972f-4c8b9bf008db" (UID: "788b1631-a897-4929-972f-4c8b9bf008db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:30 crc kubenswrapper[4828]: I1210 19:25:30.981459 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-config-data" (OuterVolumeSpecName: "config-data") pod "788b1631-a897-4929-972f-4c8b9bf008db" (UID: "788b1631-a897-4929-972f-4c8b9bf008db"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.045701 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-repo-setup-combined-ca-bundle\") pod \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\" (UID: \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\") " Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.045915 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-inventory\") pod \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\" (UID: \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\") " Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.045952 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-ssh-key\") pod \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\" (UID: \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\") " Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.046028 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtkzg\" (UniqueName: \"kubernetes.io/projected/bd51b75d-403a-4b14-8f0f-72cd8d989bca-kube-api-access-wtkzg\") pod \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\" (UID: \"bd51b75d-403a-4b14-8f0f-72cd8d989bca\") " Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.046603 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.046621 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjn9x\" (UniqueName: \"kubernetes.io/projected/788b1631-a897-4929-972f-4c8b9bf008db-kube-api-access-kjn9x\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.046633 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.046642 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/788b1631-a897-4929-972f-4c8b9bf008db-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.050826 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd51b75d-403a-4b14-8f0f-72cd8d989bca-kube-api-access-wtkzg" (OuterVolumeSpecName: "kube-api-access-wtkzg") pod "bd51b75d-403a-4b14-8f0f-72cd8d989bca" (UID: "bd51b75d-403a-4b14-8f0f-72cd8d989bca"). InnerVolumeSpecName "kube-api-access-wtkzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.051307 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "bd51b75d-403a-4b14-8f0f-72cd8d989bca" (UID: "bd51b75d-403a-4b14-8f0f-72cd8d989bca"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.078988 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-inventory" (OuterVolumeSpecName: "inventory") pod "bd51b75d-403a-4b14-8f0f-72cd8d989bca" (UID: "bd51b75d-403a-4b14-8f0f-72cd8d989bca"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.084151 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bd51b75d-403a-4b14-8f0f-72cd8d989bca" (UID: "bd51b75d-403a-4b14-8f0f-72cd8d989bca"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.148908 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtkzg\" (UniqueName: \"kubernetes.io/projected/bd51b75d-403a-4b14-8f0f-72cd8d989bca-kube-api-access-wtkzg\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.148951 4828 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.148962 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.148972 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd51b75d-403a-4b14-8f0f-72cd8d989bca-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.397468 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" event={"ID":"bd51b75d-403a-4b14-8f0f-72cd8d989bca","Type":"ContainerDied","Data":"600503f7ce8f6d7620ea5b5f3cb0bad892f750d4dacd29165a784fad4e9aa44f"} Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.397500 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.397524 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="600503f7ce8f6d7620ea5b5f3cb0bad892f750d4dacd29165a784fad4e9aa44f" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.400271 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-7krk4" event={"ID":"788b1631-a897-4929-972f-4c8b9bf008db","Type":"ContainerDied","Data":"d7b21c0329861f4f2b3afbeb34739951ae9ef6a4d3db6b450945c4917ebccc35"} Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.400303 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7b21c0329861f4f2b3afbeb34739951ae9ef6a4d3db6b450945c4917ebccc35" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.400350 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-7krk4" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.525991 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4"] Dec 10 19:25:31 crc kubenswrapper[4828]: E1210 19:25:31.526524 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788b1631-a897-4929-972f-4c8b9bf008db" containerName="aodh-db-sync" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.526540 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="788b1631-a897-4929-972f-4c8b9bf008db" containerName="aodh-db-sync" Dec 10 19:25:31 crc kubenswrapper[4828]: E1210 19:25:31.526560 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b" containerName="heat-engine" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.526566 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b" containerName="heat-engine" Dec 10 19:25:31 crc kubenswrapper[4828]: E1210 19:25:31.526592 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd51b75d-403a-4b14-8f0f-72cd8d989bca" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.526601 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd51b75d-403a-4b14-8f0f-72cd8d989bca" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.526837 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd51b75d-403a-4b14-8f0f-72cd8d989bca" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.526858 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="788b1631-a897-4929-972f-4c8b9bf008db" containerName="aodh-db-sync" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.526872 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b" containerName="heat-engine" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.527646 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.533006 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.533088 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.533338 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.537269 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.541098 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4"] Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.661084 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/12dcc39f-0cfe-498a-8de3-459c73544b77-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8psc4\" (UID: \"12dcc39f-0cfe-498a-8de3-459c73544b77\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.661276 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12dcc39f-0cfe-498a-8de3-459c73544b77-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8psc4\" (UID: \"12dcc39f-0cfe-498a-8de3-459c73544b77\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.661336 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl57g\" (UniqueName: \"kubernetes.io/projected/12dcc39f-0cfe-498a-8de3-459c73544b77-kube-api-access-kl57g\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8psc4\" (UID: \"12dcc39f-0cfe-498a-8de3-459c73544b77\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.763871 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/12dcc39f-0cfe-498a-8de3-459c73544b77-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8psc4\" (UID: \"12dcc39f-0cfe-498a-8de3-459c73544b77\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.764008 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12dcc39f-0cfe-498a-8de3-459c73544b77-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8psc4\" (UID: \"12dcc39f-0cfe-498a-8de3-459c73544b77\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.764049 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl57g\" (UniqueName: \"kubernetes.io/projected/12dcc39f-0cfe-498a-8de3-459c73544b77-kube-api-access-kl57g\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8psc4\" (UID: \"12dcc39f-0cfe-498a-8de3-459c73544b77\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.767982 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12dcc39f-0cfe-498a-8de3-459c73544b77-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8psc4\" (UID: \"12dcc39f-0cfe-498a-8de3-459c73544b77\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.774282 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/12dcc39f-0cfe-498a-8de3-459c73544b77-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8psc4\" (UID: \"12dcc39f-0cfe-498a-8de3-459c73544b77\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.779832 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl57g\" (UniqueName: \"kubernetes.io/projected/12dcc39f-0cfe-498a-8de3-459c73544b77-kube-api-access-kl57g\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8psc4\" (UID: \"12dcc39f-0cfe-498a-8de3-459c73544b77\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.801411 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b" path="/var/lib/kubelet/pods/7bb7c412-a3e7-40c9-9b1a-7432f87e5d3b/volumes" Dec 10 19:25:31 crc kubenswrapper[4828]: I1210 19:25:31.847514 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" Dec 10 19:25:32 crc kubenswrapper[4828]: I1210 19:25:32.358664 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4"] Dec 10 19:25:32 crc kubenswrapper[4828]: I1210 19:25:32.411116 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" event={"ID":"12dcc39f-0cfe-498a-8de3-459c73544b77","Type":"ContainerStarted","Data":"b6b0134619931d4d3db927a31350050253597fb4123b1cd6310fd70530895050"} Dec 10 19:25:32 crc kubenswrapper[4828]: I1210 19:25:32.789162 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:25:32 crc kubenswrapper[4828]: E1210 19:25:32.789906 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:25:33 crc kubenswrapper[4828]: I1210 19:25:33.423772 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" event={"ID":"12dcc39f-0cfe-498a-8de3-459c73544b77","Type":"ContainerStarted","Data":"1fd78d3d787a236f6b95007a2822d7ed7a79ca2b2bdac36bc34b9eb64a039e57"} Dec 10 19:25:33 crc kubenswrapper[4828]: I1210 19:25:33.443669 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" podStartSLOduration=1.786589076 podStartE2EDuration="2.443651145s" podCreationTimestamp="2025-12-10 19:25:31 +0000 UTC" firstStartedPulling="2025-12-10 19:25:32.360522185 +0000 UTC m=+1812.871133180" lastFinishedPulling="2025-12-10 19:25:33.017584254 +0000 UTC m=+1813.528195249" observedRunningTime="2025-12-10 19:25:33.439100733 +0000 UTC m=+1813.949711728" watchObservedRunningTime="2025-12-10 19:25:33.443651145 +0000 UTC m=+1813.954262150" Dec 10 19:25:36 crc kubenswrapper[4828]: I1210 19:25:36.377271 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 10 19:25:36 crc kubenswrapper[4828]: I1210 19:25:36.458702 4828 generic.go:334] "Generic (PLEG): container finished" podID="12dcc39f-0cfe-498a-8de3-459c73544b77" containerID="1fd78d3d787a236f6b95007a2822d7ed7a79ca2b2bdac36bc34b9eb64a039e57" exitCode=0 Dec 10 19:25:36 crc kubenswrapper[4828]: I1210 19:25:36.458816 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" event={"ID":"12dcc39f-0cfe-498a-8de3-459c73544b77","Type":"ContainerDied","Data":"1fd78d3d787a236f6b95007a2822d7ed7a79ca2b2bdac36bc34b9eb64a039e57"} Dec 10 19:25:36 crc kubenswrapper[4828]: I1210 19:25:36.458984 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerName="aodh-api" containerID="cri-o://82c6184925fd5fc91a8587873d8fb36c3832969f942270b28d1e662c0ac33dd9" gracePeriod=30 Dec 10 19:25:36 crc kubenswrapper[4828]: I1210 19:25:36.459004 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerName="aodh-listener" containerID="cri-o://db3ae826961c858928c3278e11bbc8cf3855efc81b2c03480a7363089c12f59d" gracePeriod=30 Dec 10 19:25:36 crc kubenswrapper[4828]: I1210 19:25:36.459059 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerName="aodh-notifier" containerID="cri-o://5a817e7054f0f1e9d6f8499822b0b209fe69c5141846f5cd362dd2014e5dfeae" gracePeriod=30 Dec 10 19:25:36 crc kubenswrapper[4828]: I1210 19:25:36.459093 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerName="aodh-evaluator" containerID="cri-o://bd19855d1153a299acb9904c79a37fc78b3e4a29e14b371d6a11db2bfff5989a" gracePeriod=30 Dec 10 19:25:37 crc kubenswrapper[4828]: I1210 19:25:37.475068 4828 generic.go:334] "Generic (PLEG): container finished" podID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerID="bd19855d1153a299acb9904c79a37fc78b3e4a29e14b371d6a11db2bfff5989a" exitCode=0 Dec 10 19:25:37 crc kubenswrapper[4828]: I1210 19:25:37.475430 4828 generic.go:334] "Generic (PLEG): container finished" podID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerID="82c6184925fd5fc91a8587873d8fb36c3832969f942270b28d1e662c0ac33dd9" exitCode=0 Dec 10 19:25:37 crc kubenswrapper[4828]: I1210 19:25:37.475159 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c3ea5ad2-3e85-405a-8e36-01578676f4e9","Type":"ContainerDied","Data":"bd19855d1153a299acb9904c79a37fc78b3e4a29e14b371d6a11db2bfff5989a"} Dec 10 19:25:37 crc kubenswrapper[4828]: I1210 19:25:37.475553 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c3ea5ad2-3e85-405a-8e36-01578676f4e9","Type":"ContainerDied","Data":"82c6184925fd5fc91a8587873d8fb36c3832969f942270b28d1e662c0ac33dd9"} Dec 10 19:25:37 crc kubenswrapper[4828]: I1210 19:25:37.959303 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.109097 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/12dcc39f-0cfe-498a-8de3-459c73544b77-ssh-key\") pod \"12dcc39f-0cfe-498a-8de3-459c73544b77\" (UID: \"12dcc39f-0cfe-498a-8de3-459c73544b77\") " Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.109288 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl57g\" (UniqueName: \"kubernetes.io/projected/12dcc39f-0cfe-498a-8de3-459c73544b77-kube-api-access-kl57g\") pod \"12dcc39f-0cfe-498a-8de3-459c73544b77\" (UID: \"12dcc39f-0cfe-498a-8de3-459c73544b77\") " Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.109467 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12dcc39f-0cfe-498a-8de3-459c73544b77-inventory\") pod \"12dcc39f-0cfe-498a-8de3-459c73544b77\" (UID: \"12dcc39f-0cfe-498a-8de3-459c73544b77\") " Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.114784 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12dcc39f-0cfe-498a-8de3-459c73544b77-kube-api-access-kl57g" (OuterVolumeSpecName: "kube-api-access-kl57g") pod "12dcc39f-0cfe-498a-8de3-459c73544b77" (UID: "12dcc39f-0cfe-498a-8de3-459c73544b77"). InnerVolumeSpecName "kube-api-access-kl57g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.140517 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12dcc39f-0cfe-498a-8de3-459c73544b77-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "12dcc39f-0cfe-498a-8de3-459c73544b77" (UID: "12dcc39f-0cfe-498a-8de3-459c73544b77"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.151553 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12dcc39f-0cfe-498a-8de3-459c73544b77-inventory" (OuterVolumeSpecName: "inventory") pod "12dcc39f-0cfe-498a-8de3-459c73544b77" (UID: "12dcc39f-0cfe-498a-8de3-459c73544b77"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.212543 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/12dcc39f-0cfe-498a-8de3-459c73544b77-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.212573 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl57g\" (UniqueName: \"kubernetes.io/projected/12dcc39f-0cfe-498a-8de3-459c73544b77-kube-api-access-kl57g\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.212584 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/12dcc39f-0cfe-498a-8de3-459c73544b77-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.496439 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" event={"ID":"12dcc39f-0cfe-498a-8de3-459c73544b77","Type":"ContainerDied","Data":"b6b0134619931d4d3db927a31350050253597fb4123b1cd6310fd70530895050"} Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.496787 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6b0134619931d4d3db927a31350050253597fb4123b1cd6310fd70530895050" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.496724 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8psc4" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.524614 4828 scope.go:117] "RemoveContainer" containerID="c1434316ba2de446c6bfb3c8f17f6eb29e3aa2954ec30915cc61ea0b8a9f2181" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.565610 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m"] Dec 10 19:25:38 crc kubenswrapper[4828]: E1210 19:25:38.566098 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12dcc39f-0cfe-498a-8de3-459c73544b77" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.566119 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="12dcc39f-0cfe-498a-8de3-459c73544b77" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.566427 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="12dcc39f-0cfe-498a-8de3-459c73544b77" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.567480 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.571851 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.572096 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.571938 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.572404 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.593599 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m"] Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.605818 4828 scope.go:117] "RemoveContainer" containerID="5a03b2ce640c72ef01f72c38f50700150bd2f332dd20556ad43d0e5d5291d8ad" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.722614 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m\" (UID: \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.722814 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj2mp\" (UniqueName: \"kubernetes.io/projected/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-kube-api-access-hj2mp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m\" (UID: \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.723114 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m\" (UID: \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.723431 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m\" (UID: \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.825575 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m\" (UID: \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.825678 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m\" (UID: \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.825743 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj2mp\" (UniqueName: \"kubernetes.io/projected/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-kube-api-access-hj2mp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m\" (UID: \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.825882 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m\" (UID: \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.830302 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m\" (UID: \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.830654 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m\" (UID: \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.830696 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m\" (UID: \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.845407 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj2mp\" (UniqueName: \"kubernetes.io/projected/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-kube-api-access-hj2mp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m\" (UID: \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" Dec 10 19:25:38 crc kubenswrapper[4828]: I1210 19:25:38.915116 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" Dec 10 19:25:39 crc kubenswrapper[4828]: I1210 19:25:39.483100 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m"] Dec 10 19:25:39 crc kubenswrapper[4828]: W1210 19:25:39.484393 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc937bc31_7116_4ab2_bd4e_c5e2bfd10e69.slice/crio-7839adf8c09a0e692fc8a8f19abab2b501028b29c9c62224400da034688e7df1 WatchSource:0}: Error finding container 7839adf8c09a0e692fc8a8f19abab2b501028b29c9c62224400da034688e7df1: Status 404 returned error can't find the container with id 7839adf8c09a0e692fc8a8f19abab2b501028b29c9c62224400da034688e7df1 Dec 10 19:25:39 crc kubenswrapper[4828]: I1210 19:25:39.520240 4828 generic.go:334] "Generic (PLEG): container finished" podID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerID="db3ae826961c858928c3278e11bbc8cf3855efc81b2c03480a7363089c12f59d" exitCode=0 Dec 10 19:25:39 crc kubenswrapper[4828]: I1210 19:25:39.520317 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c3ea5ad2-3e85-405a-8e36-01578676f4e9","Type":"ContainerDied","Data":"db3ae826961c858928c3278e11bbc8cf3855efc81b2c03480a7363089c12f59d"} Dec 10 19:25:39 crc kubenswrapper[4828]: I1210 19:25:39.521841 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" event={"ID":"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69","Type":"ContainerStarted","Data":"7839adf8c09a0e692fc8a8f19abab2b501028b29c9c62224400da034688e7df1"} Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.074101 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.158494 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-config-data\") pod \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.158954 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-internal-tls-certs\") pod \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.159057 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-public-tls-certs\") pod \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.159135 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xptc\" (UniqueName: \"kubernetes.io/projected/c3ea5ad2-3e85-405a-8e36-01578676f4e9-kube-api-access-4xptc\") pod \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.159197 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-combined-ca-bundle\") pod \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.159263 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-scripts\") pod \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.165608 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3ea5ad2-3e85-405a-8e36-01578676f4e9-kube-api-access-4xptc" (OuterVolumeSpecName: "kube-api-access-4xptc") pod "c3ea5ad2-3e85-405a-8e36-01578676f4e9" (UID: "c3ea5ad2-3e85-405a-8e36-01578676f4e9"). InnerVolumeSpecName "kube-api-access-4xptc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.185932 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-scripts" (OuterVolumeSpecName: "scripts") pod "c3ea5ad2-3e85-405a-8e36-01578676f4e9" (UID: "c3ea5ad2-3e85-405a-8e36-01578676f4e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.261405 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c3ea5ad2-3e85-405a-8e36-01578676f4e9" (UID: "c3ea5ad2-3e85-405a-8e36-01578676f4e9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.262505 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-public-tls-certs\") pod \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\" (UID: \"c3ea5ad2-3e85-405a-8e36-01578676f4e9\") " Dec 10 19:25:40 crc kubenswrapper[4828]: W1210 19:25:40.263014 4828 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/c3ea5ad2-3e85-405a-8e36-01578676f4e9/volumes/kubernetes.io~secret/public-tls-certs Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.263027 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c3ea5ad2-3e85-405a-8e36-01578676f4e9" (UID: "c3ea5ad2-3e85-405a-8e36-01578676f4e9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.264224 4828 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-scripts\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.264247 4828 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.264262 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xptc\" (UniqueName: \"kubernetes.io/projected/c3ea5ad2-3e85-405a-8e36-01578676f4e9-kube-api-access-4xptc\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.337355 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c3ea5ad2-3e85-405a-8e36-01578676f4e9" (UID: "c3ea5ad2-3e85-405a-8e36-01578676f4e9"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.367178 4828 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.402712 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-config-data" (OuterVolumeSpecName: "config-data") pod "c3ea5ad2-3e85-405a-8e36-01578676f4e9" (UID: "c3ea5ad2-3e85-405a-8e36-01578676f4e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.403178 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3ea5ad2-3e85-405a-8e36-01578676f4e9" (UID: "c3ea5ad2-3e85-405a-8e36-01578676f4e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.469872 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.469913 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3ea5ad2-3e85-405a-8e36-01578676f4e9-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.533405 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" event={"ID":"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69","Type":"ContainerStarted","Data":"71bc08eb21951337a8840768b8055e04af0b7aa9220c896962f5eaddc37a8651"} Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.554278 4828 generic.go:334] "Generic (PLEG): container finished" podID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerID="5a817e7054f0f1e9d6f8499822b0b209fe69c5141846f5cd362dd2014e5dfeae" exitCode=0 Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.554326 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c3ea5ad2-3e85-405a-8e36-01578676f4e9","Type":"ContainerDied","Data":"5a817e7054f0f1e9d6f8499822b0b209fe69c5141846f5cd362dd2014e5dfeae"} Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.554356 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c3ea5ad2-3e85-405a-8e36-01578676f4e9","Type":"ContainerDied","Data":"24cc3143556092f8e7df8f375a3fad56a976d0cd66ad6fa849e2888b713df707"} Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.554371 4828 scope.go:117] "RemoveContainer" containerID="db3ae826961c858928c3278e11bbc8cf3855efc81b2c03480a7363089c12f59d" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.554568 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.563820 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" podStartSLOduration=1.941376907 podStartE2EDuration="2.563787677s" podCreationTimestamp="2025-12-10 19:25:38 +0000 UTC" firstStartedPulling="2025-12-10 19:25:39.486693478 +0000 UTC m=+1819.997304483" lastFinishedPulling="2025-12-10 19:25:40.109104238 +0000 UTC m=+1820.619715253" observedRunningTime="2025-12-10 19:25:40.560878419 +0000 UTC m=+1821.071489424" watchObservedRunningTime="2025-12-10 19:25:40.563787677 +0000 UTC m=+1821.074398682" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.612846 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.626641 4828 scope.go:117] "RemoveContainer" containerID="5a817e7054f0f1e9d6f8499822b0b209fe69c5141846f5cd362dd2014e5dfeae" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.629739 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.658467 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 10 19:25:40 crc kubenswrapper[4828]: E1210 19:25:40.661993 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerName="aodh-listener" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.662019 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerName="aodh-listener" Dec 10 19:25:40 crc kubenswrapper[4828]: E1210 19:25:40.662032 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerName="aodh-notifier" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.662040 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerName="aodh-notifier" Dec 10 19:25:40 crc kubenswrapper[4828]: E1210 19:25:40.662068 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerName="aodh-api" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.662074 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerName="aodh-api" Dec 10 19:25:40 crc kubenswrapper[4828]: E1210 19:25:40.662091 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerName="aodh-evaluator" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.662097 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerName="aodh-evaluator" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.662331 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerName="aodh-evaluator" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.662348 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerName="aodh-listener" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.662371 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerName="aodh-api" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.662381 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" containerName="aodh-notifier" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.667396 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.672105 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.675937 4828 scope.go:117] "RemoveContainer" containerID="bd19855d1153a299acb9904c79a37fc78b3e4a29e14b371d6a11db2bfff5989a" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.676487 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.676669 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-qg8rx" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.676877 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.678520 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.678892 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.735906 4828 scope.go:117] "RemoveContainer" containerID="82c6184925fd5fc91a8587873d8fb36c3832969f942270b28d1e662c0ac33dd9" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.768224 4828 scope.go:117] "RemoveContainer" containerID="db3ae826961c858928c3278e11bbc8cf3855efc81b2c03480a7363089c12f59d" Dec 10 19:25:40 crc kubenswrapper[4828]: E1210 19:25:40.768734 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db3ae826961c858928c3278e11bbc8cf3855efc81b2c03480a7363089c12f59d\": container with ID starting with db3ae826961c858928c3278e11bbc8cf3855efc81b2c03480a7363089c12f59d not found: ID does not exist" containerID="db3ae826961c858928c3278e11bbc8cf3855efc81b2c03480a7363089c12f59d" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.768787 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db3ae826961c858928c3278e11bbc8cf3855efc81b2c03480a7363089c12f59d"} err="failed to get container status \"db3ae826961c858928c3278e11bbc8cf3855efc81b2c03480a7363089c12f59d\": rpc error: code = NotFound desc = could not find container \"db3ae826961c858928c3278e11bbc8cf3855efc81b2c03480a7363089c12f59d\": container with ID starting with db3ae826961c858928c3278e11bbc8cf3855efc81b2c03480a7363089c12f59d not found: ID does not exist" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.768832 4828 scope.go:117] "RemoveContainer" containerID="5a817e7054f0f1e9d6f8499822b0b209fe69c5141846f5cd362dd2014e5dfeae" Dec 10 19:25:40 crc kubenswrapper[4828]: E1210 19:25:40.769195 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a817e7054f0f1e9d6f8499822b0b209fe69c5141846f5cd362dd2014e5dfeae\": container with ID starting with 5a817e7054f0f1e9d6f8499822b0b209fe69c5141846f5cd362dd2014e5dfeae not found: ID does not exist" containerID="5a817e7054f0f1e9d6f8499822b0b209fe69c5141846f5cd362dd2014e5dfeae" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.769232 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a817e7054f0f1e9d6f8499822b0b209fe69c5141846f5cd362dd2014e5dfeae"} err="failed to get container status \"5a817e7054f0f1e9d6f8499822b0b209fe69c5141846f5cd362dd2014e5dfeae\": rpc error: code = NotFound desc = could not find container \"5a817e7054f0f1e9d6f8499822b0b209fe69c5141846f5cd362dd2014e5dfeae\": container with ID starting with 5a817e7054f0f1e9d6f8499822b0b209fe69c5141846f5cd362dd2014e5dfeae not found: ID does not exist" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.769249 4828 scope.go:117] "RemoveContainer" containerID="bd19855d1153a299acb9904c79a37fc78b3e4a29e14b371d6a11db2bfff5989a" Dec 10 19:25:40 crc kubenswrapper[4828]: E1210 19:25:40.769475 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd19855d1153a299acb9904c79a37fc78b3e4a29e14b371d6a11db2bfff5989a\": container with ID starting with bd19855d1153a299acb9904c79a37fc78b3e4a29e14b371d6a11db2bfff5989a not found: ID does not exist" containerID="bd19855d1153a299acb9904c79a37fc78b3e4a29e14b371d6a11db2bfff5989a" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.769548 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd19855d1153a299acb9904c79a37fc78b3e4a29e14b371d6a11db2bfff5989a"} err="failed to get container status \"bd19855d1153a299acb9904c79a37fc78b3e4a29e14b371d6a11db2bfff5989a\": rpc error: code = NotFound desc = could not find container \"bd19855d1153a299acb9904c79a37fc78b3e4a29e14b371d6a11db2bfff5989a\": container with ID starting with bd19855d1153a299acb9904c79a37fc78b3e4a29e14b371d6a11db2bfff5989a not found: ID does not exist" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.769586 4828 scope.go:117] "RemoveContainer" containerID="82c6184925fd5fc91a8587873d8fb36c3832969f942270b28d1e662c0ac33dd9" Dec 10 19:25:40 crc kubenswrapper[4828]: E1210 19:25:40.769982 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82c6184925fd5fc91a8587873d8fb36c3832969f942270b28d1e662c0ac33dd9\": container with ID starting with 82c6184925fd5fc91a8587873d8fb36c3832969f942270b28d1e662c0ac33dd9 not found: ID does not exist" containerID="82c6184925fd5fc91a8587873d8fb36c3832969f942270b28d1e662c0ac33dd9" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.770022 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82c6184925fd5fc91a8587873d8fb36c3832969f942270b28d1e662c0ac33dd9"} err="failed to get container status \"82c6184925fd5fc91a8587873d8fb36c3832969f942270b28d1e662c0ac33dd9\": rpc error: code = NotFound desc = could not find container \"82c6184925fd5fc91a8587873d8fb36c3832969f942270b28d1e662c0ac33dd9\": container with ID starting with 82c6184925fd5fc91a8587873d8fb36c3832969f942270b28d1e662c0ac33dd9 not found: ID does not exist" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.780157 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-internal-tls-certs\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.780212 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-config-data\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.780229 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-combined-ca-bundle\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.780297 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh2hv\" (UniqueName: \"kubernetes.io/projected/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-kube-api-access-sh2hv\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.780443 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-scripts\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.780463 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-public-tls-certs\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.882582 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-scripts\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.882621 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-public-tls-certs\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.882809 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-internal-tls-certs\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.882845 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-config-data\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.882862 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-combined-ca-bundle\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.882886 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh2hv\" (UniqueName: \"kubernetes.io/projected/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-kube-api-access-sh2hv\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.887240 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-internal-tls-certs\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.887255 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-config-data\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.887588 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-combined-ca-bundle\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.887696 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-public-tls-certs\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.889632 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-scripts\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:40 crc kubenswrapper[4828]: I1210 19:25:40.903408 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh2hv\" (UniqueName: \"kubernetes.io/projected/683b1cf4-b6dc-4bf5-a537-6461533fc1ef-kube-api-access-sh2hv\") pod \"aodh-0\" (UID: \"683b1cf4-b6dc-4bf5-a537-6461533fc1ef\") " pod="openstack/aodh-0" Dec 10 19:25:41 crc kubenswrapper[4828]: I1210 19:25:41.018515 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 10 19:25:41 crc kubenswrapper[4828]: I1210 19:25:41.499482 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 10 19:25:41 crc kubenswrapper[4828]: W1210 19:25:41.499697 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod683b1cf4_b6dc_4bf5_a537_6461533fc1ef.slice/crio-aa9b8c824ba738f2d6a4d963054c16318d9446e48361894e6374a23d615bd6d1 WatchSource:0}: Error finding container aa9b8c824ba738f2d6a4d963054c16318d9446e48361894e6374a23d615bd6d1: Status 404 returned error can't find the container with id aa9b8c824ba738f2d6a4d963054c16318d9446e48361894e6374a23d615bd6d1 Dec 10 19:25:41 crc kubenswrapper[4828]: I1210 19:25:41.580869 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"683b1cf4-b6dc-4bf5-a537-6461533fc1ef","Type":"ContainerStarted","Data":"aa9b8c824ba738f2d6a4d963054c16318d9446e48361894e6374a23d615bd6d1"} Dec 10 19:25:41 crc kubenswrapper[4828]: I1210 19:25:41.813217 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3ea5ad2-3e85-405a-8e36-01578676f4e9" path="/var/lib/kubelet/pods/c3ea5ad2-3e85-405a-8e36-01578676f4e9/volumes" Dec 10 19:25:42 crc kubenswrapper[4828]: I1210 19:25:42.599841 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"683b1cf4-b6dc-4bf5-a537-6461533fc1ef","Type":"ContainerStarted","Data":"8e76b0331819dc8d7ef0df0df913cb47d944825432691d60e780c32c2b54d0cf"} Dec 10 19:25:44 crc kubenswrapper[4828]: I1210 19:25:44.626316 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"683b1cf4-b6dc-4bf5-a537-6461533fc1ef","Type":"ContainerStarted","Data":"7488be241e922bf10e3b8eaa529c5292d569b9c05fffe015c058139c0cdfa11a"} Dec 10 19:25:44 crc kubenswrapper[4828]: I1210 19:25:44.789235 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:25:44 crc kubenswrapper[4828]: E1210 19:25:44.789516 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:25:45 crc kubenswrapper[4828]: I1210 19:25:45.648881 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"683b1cf4-b6dc-4bf5-a537-6461533fc1ef","Type":"ContainerStarted","Data":"2a71c429c0f58be7c93b9c968b8330bb5e63848786a45cfa481aa9fc0837fe1c"} Dec 10 19:25:47 crc kubenswrapper[4828]: I1210 19:25:47.680631 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"683b1cf4-b6dc-4bf5-a537-6461533fc1ef","Type":"ContainerStarted","Data":"aab3f5ef8a2a54fb370cdb1b07846437fae659f07416d553dc4ed0f84b25375c"} Dec 10 19:25:56 crc kubenswrapper[4828]: I1210 19:25:56.789213 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:25:57 crc kubenswrapper[4828]: I1210 19:25:57.804371 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"b4415f07501c6dc89d7e4fe4d4e51bb63a0dedd7e1962e15bc82fe0414c7a987"} Dec 10 19:25:57 crc kubenswrapper[4828]: I1210 19:25:57.830220 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=12.201245386 podStartE2EDuration="17.83020415s" podCreationTimestamp="2025-12-10 19:25:40 +0000 UTC" firstStartedPulling="2025-12-10 19:25:41.504170747 +0000 UTC m=+1822.014781752" lastFinishedPulling="2025-12-10 19:25:47.133129511 +0000 UTC m=+1827.643740516" observedRunningTime="2025-12-10 19:25:47.704445029 +0000 UTC m=+1828.215056034" watchObservedRunningTime="2025-12-10 19:25:57.83020415 +0000 UTC m=+1838.340815155" Dec 10 19:26:38 crc kubenswrapper[4828]: I1210 19:26:38.838146 4828 scope.go:117] "RemoveContainer" containerID="a416cfc1250e09a7ae12693c3a210bf1319994086d1e3b12d3e1032bd32df10d" Dec 10 19:27:38 crc kubenswrapper[4828]: I1210 19:27:38.922909 4828 scope.go:117] "RemoveContainer" containerID="7761a7fab7f1757991c79aa7082709c281bce53ed2c6d4d94fb4ef159ff329b9" Dec 10 19:27:38 crc kubenswrapper[4828]: I1210 19:27:38.960168 4828 scope.go:117] "RemoveContainer" containerID="3f7f00335c86a1e5b99e0e3677a9c6cf168454f042fc2f5a720193b77fff2c7e" Dec 10 19:28:21 crc kubenswrapper[4828]: I1210 19:28:21.230718 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:28:21 crc kubenswrapper[4828]: I1210 19:28:21.231314 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:28:39 crc kubenswrapper[4828]: I1210 19:28:39.091587 4828 scope.go:117] "RemoveContainer" containerID="68e5730b588bb1538a77492edf973b42c140868fae4a86ba497bcb09bbca320a" Dec 10 19:28:39 crc kubenswrapper[4828]: I1210 19:28:39.151278 4828 scope.go:117] "RemoveContainer" containerID="b25269503a6063695d2f58c6405d12d812e495ce0a08ab209d320516e49f14bf" Dec 10 19:28:39 crc kubenswrapper[4828]: I1210 19:28:39.197456 4828 scope.go:117] "RemoveContainer" containerID="bfc226dee05895e01f3129c274e91dca9ad880d14769124bb765b6acdfe395a4" Dec 10 19:28:39 crc kubenswrapper[4828]: I1210 19:28:39.223667 4828 scope.go:117] "RemoveContainer" containerID="885f248b67052747290c8d222d6f14491a48c593f5c7bc9eb5f84872ea81de9a" Dec 10 19:28:45 crc kubenswrapper[4828]: I1210 19:28:45.724391 4828 generic.go:334] "Generic (PLEG): container finished" podID="c937bc31-7116-4ab2-bd4e-c5e2bfd10e69" containerID="71bc08eb21951337a8840768b8055e04af0b7aa9220c896962f5eaddc37a8651" exitCode=0 Dec 10 19:28:45 crc kubenswrapper[4828]: I1210 19:28:45.724933 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" event={"ID":"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69","Type":"ContainerDied","Data":"71bc08eb21951337a8840768b8055e04af0b7aa9220c896962f5eaddc37a8651"} Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.194051 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.337240 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-bootstrap-combined-ca-bundle\") pod \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\" (UID: \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\") " Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.337666 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-ssh-key\") pod \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\" (UID: \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\") " Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.337845 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-inventory\") pod \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\" (UID: \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\") " Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.338084 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hj2mp\" (UniqueName: \"kubernetes.io/projected/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-kube-api-access-hj2mp\") pod \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\" (UID: \"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69\") " Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.350037 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "c937bc31-7116-4ab2-bd4e-c5e2bfd10e69" (UID: "c937bc31-7116-4ab2-bd4e-c5e2bfd10e69"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.350073 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-kube-api-access-hj2mp" (OuterVolumeSpecName: "kube-api-access-hj2mp") pod "c937bc31-7116-4ab2-bd4e-c5e2bfd10e69" (UID: "c937bc31-7116-4ab2-bd4e-c5e2bfd10e69"). InnerVolumeSpecName "kube-api-access-hj2mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.377830 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c937bc31-7116-4ab2-bd4e-c5e2bfd10e69" (UID: "c937bc31-7116-4ab2-bd4e-c5e2bfd10e69"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.385041 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-inventory" (OuterVolumeSpecName: "inventory") pod "c937bc31-7116-4ab2-bd4e-c5e2bfd10e69" (UID: "c937bc31-7116-4ab2-bd4e-c5e2bfd10e69"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.441054 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.441088 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hj2mp\" (UniqueName: \"kubernetes.io/projected/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-kube-api-access-hj2mp\") on node \"crc\" DevicePath \"\"" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.441099 4828 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.441108 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c937bc31-7116-4ab2-bd4e-c5e2bfd10e69-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.762836 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" event={"ID":"c937bc31-7116-4ab2-bd4e-c5e2bfd10e69","Type":"ContainerDied","Data":"7839adf8c09a0e692fc8a8f19abab2b501028b29c9c62224400da034688e7df1"} Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.762928 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7839adf8c09a0e692fc8a8f19abab2b501028b29c9c62224400da034688e7df1" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.762865 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.835573 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq"] Dec 10 19:28:47 crc kubenswrapper[4828]: E1210 19:28:47.836192 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c937bc31-7116-4ab2-bd4e-c5e2bfd10e69" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.836214 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c937bc31-7116-4ab2-bd4e-c5e2bfd10e69" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.836428 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c937bc31-7116-4ab2-bd4e-c5e2bfd10e69" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.837368 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.839611 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.839996 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.840990 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.847538 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.848331 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq"] Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.962053 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq\" (UID: \"98d9383c-0b06-4a8d-8d4c-1832493c3c5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.962511 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mdvs\" (UniqueName: \"kubernetes.io/projected/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-kube-api-access-2mdvs\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq\" (UID: \"98d9383c-0b06-4a8d-8d4c-1832493c3c5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" Dec 10 19:28:47 crc kubenswrapper[4828]: I1210 19:28:47.962553 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq\" (UID: \"98d9383c-0b06-4a8d-8d4c-1832493c3c5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" Dec 10 19:28:48 crc kubenswrapper[4828]: I1210 19:28:48.064541 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq\" (UID: \"98d9383c-0b06-4a8d-8d4c-1832493c3c5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" Dec 10 19:28:48 crc kubenswrapper[4828]: I1210 19:28:48.064679 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq\" (UID: \"98d9383c-0b06-4a8d-8d4c-1832493c3c5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" Dec 10 19:28:48 crc kubenswrapper[4828]: I1210 19:28:48.064782 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mdvs\" (UniqueName: \"kubernetes.io/projected/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-kube-api-access-2mdvs\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq\" (UID: \"98d9383c-0b06-4a8d-8d4c-1832493c3c5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" Dec 10 19:28:48 crc kubenswrapper[4828]: I1210 19:28:48.075341 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq\" (UID: \"98d9383c-0b06-4a8d-8d4c-1832493c3c5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" Dec 10 19:28:48 crc kubenswrapper[4828]: I1210 19:28:48.075367 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq\" (UID: \"98d9383c-0b06-4a8d-8d4c-1832493c3c5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" Dec 10 19:28:48 crc kubenswrapper[4828]: I1210 19:28:48.084484 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mdvs\" (UniqueName: \"kubernetes.io/projected/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-kube-api-access-2mdvs\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq\" (UID: \"98d9383c-0b06-4a8d-8d4c-1832493c3c5f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" Dec 10 19:28:48 crc kubenswrapper[4828]: I1210 19:28:48.165150 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" Dec 10 19:28:48 crc kubenswrapper[4828]: I1210 19:28:48.672039 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq"] Dec 10 19:28:48 crc kubenswrapper[4828]: I1210 19:28:48.677394 4828 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:28:48 crc kubenswrapper[4828]: I1210 19:28:48.774411 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" event={"ID":"98d9383c-0b06-4a8d-8d4c-1832493c3c5f","Type":"ContainerStarted","Data":"8f43073f50ac151571c7d4001354a83748b5d28556bc76d95d9ec1eda488ab7b"} Dec 10 19:28:49 crc kubenswrapper[4828]: I1210 19:28:49.784915 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" event={"ID":"98d9383c-0b06-4a8d-8d4c-1832493c3c5f","Type":"ContainerStarted","Data":"b41ce6891e3abf70029b38d2cb6e69a887ea8e687f32bbcb7e139c9ab73c7828"} Dec 10 19:28:49 crc kubenswrapper[4828]: I1210 19:28:49.819064 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" podStartSLOduration=2.231648654 podStartE2EDuration="2.819038504s" podCreationTimestamp="2025-12-10 19:28:47 +0000 UTC" firstStartedPulling="2025-12-10 19:28:48.677191448 +0000 UTC m=+2009.187802453" lastFinishedPulling="2025-12-10 19:28:49.264581288 +0000 UTC m=+2009.775192303" observedRunningTime="2025-12-10 19:28:49.80397855 +0000 UTC m=+2010.314589565" watchObservedRunningTime="2025-12-10 19:28:49.819038504 +0000 UTC m=+2010.329649509" Dec 10 19:28:51 crc kubenswrapper[4828]: I1210 19:28:51.230244 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:28:51 crc kubenswrapper[4828]: I1210 19:28:51.230550 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:28:56 crc kubenswrapper[4828]: I1210 19:28:56.061574 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-8a56-account-create-update-9dg89"] Dec 10 19:28:56 crc kubenswrapper[4828]: I1210 19:28:56.081926 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-8a56-account-create-update-9dg89"] Dec 10 19:28:56 crc kubenswrapper[4828]: I1210 19:28:56.096412 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-eb79-account-create-update-kx4qg"] Dec 10 19:28:56 crc kubenswrapper[4828]: I1210 19:28:56.110512 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0756-account-create-update-qr5d5"] Dec 10 19:28:56 crc kubenswrapper[4828]: I1210 19:28:56.128203 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-z45cg"] Dec 10 19:28:56 crc kubenswrapper[4828]: I1210 19:28:56.140284 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-eb79-account-create-update-kx4qg"] Dec 10 19:28:56 crc kubenswrapper[4828]: I1210 19:28:56.151122 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0756-account-create-update-qr5d5"] Dec 10 19:28:56 crc kubenswrapper[4828]: I1210 19:28:56.162980 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-z45cg"] Dec 10 19:28:56 crc kubenswrapper[4828]: I1210 19:28:56.175216 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-6bchn"] Dec 10 19:28:56 crc kubenswrapper[4828]: I1210 19:28:56.186431 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-6bchn"] Dec 10 19:28:57 crc kubenswrapper[4828]: I1210 19:28:57.037348 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-bec3-account-create-update-gfhnk"] Dec 10 19:28:57 crc kubenswrapper[4828]: I1210 19:28:57.052561 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-vd4h8"] Dec 10 19:28:57 crc kubenswrapper[4828]: I1210 19:28:57.064902 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-74mkr"] Dec 10 19:28:57 crc kubenswrapper[4828]: I1210 19:28:57.077080 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-74mkr"] Dec 10 19:28:57 crc kubenswrapper[4828]: I1210 19:28:57.088959 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-vd4h8"] Dec 10 19:28:57 crc kubenswrapper[4828]: I1210 19:28:57.100391 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-bec3-account-create-update-gfhnk"] Dec 10 19:28:57 crc kubenswrapper[4828]: I1210 19:28:57.802545 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="019b4ce6-57e1-42b0-be80-5c810a163d46" path="/var/lib/kubelet/pods/019b4ce6-57e1-42b0-be80-5c810a163d46/volumes" Dec 10 19:28:57 crc kubenswrapper[4828]: I1210 19:28:57.804018 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1218268b-0e3f-41d5-8e3e-cac3f9164f70" path="/var/lib/kubelet/pods/1218268b-0e3f-41d5-8e3e-cac3f9164f70/volumes" Dec 10 19:28:57 crc kubenswrapper[4828]: I1210 19:28:57.804862 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32ccef7c-5acd-4ae1-a555-8079eaa96ae9" path="/var/lib/kubelet/pods/32ccef7c-5acd-4ae1-a555-8079eaa96ae9/volumes" Dec 10 19:28:57 crc kubenswrapper[4828]: I1210 19:28:57.805686 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fc92de3-2c9e-4e10-8d89-a71ffd49523d" path="/var/lib/kubelet/pods/3fc92de3-2c9e-4e10-8d89-a71ffd49523d/volumes" Dec 10 19:28:57 crc kubenswrapper[4828]: I1210 19:28:57.807146 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66db2b53-0c7d-4165-b743-99bddafea231" path="/var/lib/kubelet/pods/66db2b53-0c7d-4165-b743-99bddafea231/volumes" Dec 10 19:28:57 crc kubenswrapper[4828]: I1210 19:28:57.808178 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a" path="/var/lib/kubelet/pods/73d03b51-b8cd-4a4a-bc09-40ea8cc7d29a/volumes" Dec 10 19:28:57 crc kubenswrapper[4828]: I1210 19:28:57.808962 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81aa6d69-1494-4790-a0b3-9e13444c17da" path="/var/lib/kubelet/pods/81aa6d69-1494-4790-a0b3-9e13444c17da/volumes" Dec 10 19:28:57 crc kubenswrapper[4828]: I1210 19:28:57.809998 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e391ca51-f39a-4718-9265-4f21f26155e5" path="/var/lib/kubelet/pods/e391ca51-f39a-4718-9265-4f21f26155e5/volumes" Dec 10 19:29:04 crc kubenswrapper[4828]: I1210 19:29:04.049736 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-7f0a-account-create-update-nh472"] Dec 10 19:29:04 crc kubenswrapper[4828]: I1210 19:29:04.063637 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m"] Dec 10 19:29:04 crc kubenswrapper[4828]: I1210 19:29:04.076296 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-6nb4m"] Dec 10 19:29:04 crc kubenswrapper[4828]: I1210 19:29:04.088201 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-7f0a-account-create-update-nh472"] Dec 10 19:29:05 crc kubenswrapper[4828]: I1210 19:29:05.802055 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecab69d1-3989-4e69-90f3-b54d4e229283" path="/var/lib/kubelet/pods/ecab69d1-3989-4e69-90f3-b54d4e229283/volumes" Dec 10 19:29:05 crc kubenswrapper[4828]: I1210 19:29:05.803015 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb2e9f6f-100c-494a-bbca-44c3ab8fc46c" path="/var/lib/kubelet/pods/fb2e9f6f-100c-494a-bbca-44c3ab8fc46c/volumes" Dec 10 19:29:21 crc kubenswrapper[4828]: I1210 19:29:21.044365 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-bmrgx"] Dec 10 19:29:21 crc kubenswrapper[4828]: I1210 19:29:21.056343 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-bmrgx"] Dec 10 19:29:21 crc kubenswrapper[4828]: I1210 19:29:21.230243 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:29:21 crc kubenswrapper[4828]: I1210 19:29:21.230308 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:29:21 crc kubenswrapper[4828]: I1210 19:29:21.230368 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 19:29:21 crc kubenswrapper[4828]: I1210 19:29:21.231422 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b4415f07501c6dc89d7e4fe4d4e51bb63a0dedd7e1962e15bc82fe0414c7a987"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:29:21 crc kubenswrapper[4828]: I1210 19:29:21.231506 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://b4415f07501c6dc89d7e4fe4d4e51bb63a0dedd7e1962e15bc82fe0414c7a987" gracePeriod=600 Dec 10 19:29:21 crc kubenswrapper[4828]: I1210 19:29:21.806208 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="489d1351-3bec-4909-a5dd-7f5e14f280d6" path="/var/lib/kubelet/pods/489d1351-3bec-4909-a5dd-7f5e14f280d6/volumes" Dec 10 19:29:22 crc kubenswrapper[4828]: I1210 19:29:22.141680 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="b4415f07501c6dc89d7e4fe4d4e51bb63a0dedd7e1962e15bc82fe0414c7a987" exitCode=0 Dec 10 19:29:22 crc kubenswrapper[4828]: I1210 19:29:22.141731 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"b4415f07501c6dc89d7e4fe4d4e51bb63a0dedd7e1962e15bc82fe0414c7a987"} Dec 10 19:29:22 crc kubenswrapper[4828]: I1210 19:29:22.142109 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6"} Dec 10 19:29:22 crc kubenswrapper[4828]: I1210 19:29:22.142139 4828 scope.go:117] "RemoveContainer" containerID="ae01ed3555c9426ea6c3bd4b634b61c47c96015ee4a7e56adf8ed5bddfd7f7ca" Dec 10 19:29:24 crc kubenswrapper[4828]: I1210 19:29:24.066551 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-bbtgm"] Dec 10 19:29:24 crc kubenswrapper[4828]: I1210 19:29:24.082987 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-zkd8f"] Dec 10 19:29:24 crc kubenswrapper[4828]: I1210 19:29:24.093635 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-bbtgm"] Dec 10 19:29:24 crc kubenswrapper[4828]: I1210 19:29:24.106300 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-4btg4"] Dec 10 19:29:24 crc kubenswrapper[4828]: I1210 19:29:24.117081 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-30e8-account-create-update-5zwls"] Dec 10 19:29:24 crc kubenswrapper[4828]: I1210 19:29:24.128048 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-d565-account-create-update-vwzdt"] Dec 10 19:29:24 crc kubenswrapper[4828]: I1210 19:29:24.137718 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-zkd8f"] Dec 10 19:29:24 crc kubenswrapper[4828]: I1210 19:29:24.153911 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-d565-account-create-update-vwzdt"] Dec 10 19:29:24 crc kubenswrapper[4828]: I1210 19:29:24.168206 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-30e8-account-create-update-5zwls"] Dec 10 19:29:24 crc kubenswrapper[4828]: I1210 19:29:24.182733 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-4btg4"] Dec 10 19:29:24 crc kubenswrapper[4828]: I1210 19:29:24.197166 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-1a9d-account-create-update-ql2pg"] Dec 10 19:29:24 crc kubenswrapper[4828]: I1210 19:29:24.234914 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-ae3c-account-create-update-v6g92"] Dec 10 19:29:24 crc kubenswrapper[4828]: I1210 19:29:24.250233 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-ae3c-account-create-update-v6g92"] Dec 10 19:29:24 crc kubenswrapper[4828]: I1210 19:29:24.261232 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-1a9d-account-create-update-ql2pg"] Dec 10 19:29:25 crc kubenswrapper[4828]: I1210 19:29:25.801037 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20f5475c-6d07-4be3-96f5-fed87f4ad62a" path="/var/lib/kubelet/pods/20f5475c-6d07-4be3-96f5-fed87f4ad62a/volumes" Dec 10 19:29:25 crc kubenswrapper[4828]: I1210 19:29:25.802442 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3" path="/var/lib/kubelet/pods/37a44be7-2c0c-45d6-b8ad-3d72e0dfdfa3/volumes" Dec 10 19:29:25 crc kubenswrapper[4828]: I1210 19:29:25.803131 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43f54580-b31f-45db-b046-3fb45235c384" path="/var/lib/kubelet/pods/43f54580-b31f-45db-b046-3fb45235c384/volumes" Dec 10 19:29:25 crc kubenswrapper[4828]: I1210 19:29:25.804177 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="515a9b7f-6551-4a40-bd25-6c851129a6d0" path="/var/lib/kubelet/pods/515a9b7f-6551-4a40-bd25-6c851129a6d0/volumes" Dec 10 19:29:25 crc kubenswrapper[4828]: I1210 19:29:25.805443 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774" path="/var/lib/kubelet/pods/a8d65c41-59f4-4aa5-8c7f-b4c1a1bef774/volumes" Dec 10 19:29:25 crc kubenswrapper[4828]: I1210 19:29:25.806118 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca06dcbb-f357-454e-afd5-7bd5b42fb6ed" path="/var/lib/kubelet/pods/ca06dcbb-f357-454e-afd5-7bd5b42fb6ed/volumes" Dec 10 19:29:25 crc kubenswrapper[4828]: I1210 19:29:25.806676 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cea49d65-9ca3-41a3-b970-25b0ccaac138" path="/var/lib/kubelet/pods/cea49d65-9ca3-41a3-b970-25b0ccaac138/volumes" Dec 10 19:29:27 crc kubenswrapper[4828]: I1210 19:29:27.040938 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-9rgqj"] Dec 10 19:29:27 crc kubenswrapper[4828]: I1210 19:29:27.056127 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-9rgqj"] Dec 10 19:29:27 crc kubenswrapper[4828]: I1210 19:29:27.802923 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2" path="/var/lib/kubelet/pods/5f1645fa-ceff-4dcf-9b7e-e7b983a6ead2/volumes" Dec 10 19:29:29 crc kubenswrapper[4828]: I1210 19:29:29.036006 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-b2rhf"] Dec 10 19:29:29 crc kubenswrapper[4828]: I1210 19:29:29.047564 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-b2rhf"] Dec 10 19:29:29 crc kubenswrapper[4828]: I1210 19:29:29.803338 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="611d78ee-8203-4a4f-bb01-eedd92ab42fd" path="/var/lib/kubelet/pods/611d78ee-8203-4a4f-bb01-eedd92ab42fd/volumes" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.300041 4828 scope.go:117] "RemoveContainer" containerID="b5590164ad17189bd48f55e46455646b366b4df0533a870dcaafbc1917e0cc96" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.331870 4828 scope.go:117] "RemoveContainer" containerID="9bd8703eac224ae862977acd7a3bf7bb03820f910fd20187da8b777b36435ddc" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.394069 4828 scope.go:117] "RemoveContainer" containerID="b7bf385ea061afcbcde12b0252ddb8b55001e1197495dfa1b2b45e8107778b13" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.454150 4828 scope.go:117] "RemoveContainer" containerID="88c45d03b97d34cdd7dab5b7b715fb5e59cf4330c07fa76ef60ff0c465365dc8" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.506242 4828 scope.go:117] "RemoveContainer" containerID="bc200df36df7c83b3e98bfd8302854d7b057f40b4b3c3aff31ae44ecb938792a" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.560530 4828 scope.go:117] "RemoveContainer" containerID="6cb044f2dfd191d9d7014f07cf4aa79d64614a633efcab0bf69681601208db5e" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.585505 4828 scope.go:117] "RemoveContainer" containerID="6d3ae3c549af589ec6acd20f78c025c3bd7121f7dfd42814870195fbffa23d55" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.606267 4828 scope.go:117] "RemoveContainer" containerID="ec108eeff2ebf2be4a5e63b9f06ab16ac0c30fb65edf59a4dbf7f0bb24d8cf97" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.663886 4828 scope.go:117] "RemoveContainer" containerID="5a669fd9dee5beaa44c3ac23322606f4e2ad87a45eca78cc6e7f974d677c7b6d" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.686622 4828 scope.go:117] "RemoveContainer" containerID="89bf1a217038cec0da206b74444329ac2674155ea21be96652d74c3feddb3d09" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.708781 4828 scope.go:117] "RemoveContainer" containerID="48aff4c4f9de4966bd0b4dd353d7c334a79534b9bba9d82347bb061c5ff4b2b1" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.731148 4828 scope.go:117] "RemoveContainer" containerID="c5d601775a8e90f10c6f5da090a2eac53127cbf90d0380f63d9f9454b0eccfee" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.754460 4828 scope.go:117] "RemoveContainer" containerID="d869cd6461234fc218a1eaf175e3888c1eee95293ecc651d9ba024736d5d2c2e" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.775304 4828 scope.go:117] "RemoveContainer" containerID="2a7168e8a3b73a3f90371d4052dccae3569a9f926e27b7fd2e18b2a8a57b63f4" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.797471 4828 scope.go:117] "RemoveContainer" containerID="ecd68268f1cafc889845b4718d6959187049f757ea6e70219c5c2ca365c91efc" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.826645 4828 scope.go:117] "RemoveContainer" containerID="92c5c70edf42b6537b5effa2035053b811de539786aae6d531025d8b6ec7eb9c" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.860141 4828 scope.go:117] "RemoveContainer" containerID="bed6417b33c47d4233837e21e57c382f5464419c78faf437b09a3367f3779dee" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.882205 4828 scope.go:117] "RemoveContainer" containerID="0fea48dccbb4ec05cb44a530fb12900f0c322976b667d19f0ed7a0cba015e4a0" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.902615 4828 scope.go:117] "RemoveContainer" containerID="e3ff992a8c33d465274becc00390c4ea69eb05aa6d8dc109e5e8b2cf8bbf36b7" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.922962 4828 scope.go:117] "RemoveContainer" containerID="5697f45ea9e4d482df1cb71f59c4405623f3d5daa5661c7ac75622b2e1e2f4e7" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.943277 4828 scope.go:117] "RemoveContainer" containerID="4e3dd83d7be744622f33d7ee0d8f4daab774a0d444472c78b8b37f48caf17bde" Dec 10 19:29:39 crc kubenswrapper[4828]: I1210 19:29:39.966854 4828 scope.go:117] "RemoveContainer" containerID="a0d7410b9dd4c96e017faa103058fdfd8d45257a1322ac9f2e0749cf50a1a584" Dec 10 19:30:00 crc kubenswrapper[4828]: I1210 19:30:00.149781 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj"] Dec 10 19:30:00 crc kubenswrapper[4828]: I1210 19:30:00.152290 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj" Dec 10 19:30:00 crc kubenswrapper[4828]: I1210 19:30:00.154120 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 19:30:00 crc kubenswrapper[4828]: I1210 19:30:00.154713 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 19:30:00 crc kubenswrapper[4828]: I1210 19:30:00.167762 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj"] Dec 10 19:30:00 crc kubenswrapper[4828]: I1210 19:30:00.324542 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-secret-volume\") pod \"collect-profiles-29423250-l5gsj\" (UID: \"e8499d18-c5eb-4373-ab8d-0a20eab63bcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj" Dec 10 19:30:00 crc kubenswrapper[4828]: I1210 19:30:00.324598 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-config-volume\") pod \"collect-profiles-29423250-l5gsj\" (UID: \"e8499d18-c5eb-4373-ab8d-0a20eab63bcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj" Dec 10 19:30:00 crc kubenswrapper[4828]: I1210 19:30:00.324643 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5hms\" (UniqueName: \"kubernetes.io/projected/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-kube-api-access-q5hms\") pod \"collect-profiles-29423250-l5gsj\" (UID: \"e8499d18-c5eb-4373-ab8d-0a20eab63bcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj" Dec 10 19:30:00 crc kubenswrapper[4828]: I1210 19:30:00.426552 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-secret-volume\") pod \"collect-profiles-29423250-l5gsj\" (UID: \"e8499d18-c5eb-4373-ab8d-0a20eab63bcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj" Dec 10 19:30:00 crc kubenswrapper[4828]: I1210 19:30:00.426620 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-config-volume\") pod \"collect-profiles-29423250-l5gsj\" (UID: \"e8499d18-c5eb-4373-ab8d-0a20eab63bcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj" Dec 10 19:30:00 crc kubenswrapper[4828]: I1210 19:30:00.426695 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5hms\" (UniqueName: \"kubernetes.io/projected/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-kube-api-access-q5hms\") pod \"collect-profiles-29423250-l5gsj\" (UID: \"e8499d18-c5eb-4373-ab8d-0a20eab63bcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj" Dec 10 19:30:00 crc kubenswrapper[4828]: I1210 19:30:00.427837 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-config-volume\") pod \"collect-profiles-29423250-l5gsj\" (UID: \"e8499d18-c5eb-4373-ab8d-0a20eab63bcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj" Dec 10 19:30:00 crc kubenswrapper[4828]: I1210 19:30:00.432709 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-secret-volume\") pod \"collect-profiles-29423250-l5gsj\" (UID: \"e8499d18-c5eb-4373-ab8d-0a20eab63bcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj" Dec 10 19:30:00 crc kubenswrapper[4828]: I1210 19:30:00.448229 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5hms\" (UniqueName: \"kubernetes.io/projected/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-kube-api-access-q5hms\") pod \"collect-profiles-29423250-l5gsj\" (UID: \"e8499d18-c5eb-4373-ab8d-0a20eab63bcc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj" Dec 10 19:30:00 crc kubenswrapper[4828]: I1210 19:30:00.478617 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj" Dec 10 19:30:00 crc kubenswrapper[4828]: I1210 19:30:00.963688 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj"] Dec 10 19:30:00 crc kubenswrapper[4828]: W1210 19:30:00.973500 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8499d18_c5eb_4373_ab8d_0a20eab63bcc.slice/crio-bc461209f1b3ccae625611bd36d2b02a5f2d2d830604164a13c656424572a999 WatchSource:0}: Error finding container bc461209f1b3ccae625611bd36d2b02a5f2d2d830604164a13c656424572a999: Status 404 returned error can't find the container with id bc461209f1b3ccae625611bd36d2b02a5f2d2d830604164a13c656424572a999 Dec 10 19:30:01 crc kubenswrapper[4828]: I1210 19:30:01.569591 4828 generic.go:334] "Generic (PLEG): container finished" podID="e8499d18-c5eb-4373-ab8d-0a20eab63bcc" containerID="6a3d61247abe6456e973ec8131c3bd05e29a51b81157111c3bd315a5a00297eb" exitCode=0 Dec 10 19:30:01 crc kubenswrapper[4828]: I1210 19:30:01.569702 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj" event={"ID":"e8499d18-c5eb-4373-ab8d-0a20eab63bcc","Type":"ContainerDied","Data":"6a3d61247abe6456e973ec8131c3bd05e29a51b81157111c3bd315a5a00297eb"} Dec 10 19:30:01 crc kubenswrapper[4828]: I1210 19:30:01.570054 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj" event={"ID":"e8499d18-c5eb-4373-ab8d-0a20eab63bcc","Type":"ContainerStarted","Data":"bc461209f1b3ccae625611bd36d2b02a5f2d2d830604164a13c656424572a999"} Dec 10 19:30:02 crc kubenswrapper[4828]: I1210 19:30:02.962151 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj" Dec 10 19:30:03 crc kubenswrapper[4828]: I1210 19:30:03.103644 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-config-volume\") pod \"e8499d18-c5eb-4373-ab8d-0a20eab63bcc\" (UID: \"e8499d18-c5eb-4373-ab8d-0a20eab63bcc\") " Dec 10 19:30:03 crc kubenswrapper[4828]: I1210 19:30:03.104089 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-secret-volume\") pod \"e8499d18-c5eb-4373-ab8d-0a20eab63bcc\" (UID: \"e8499d18-c5eb-4373-ab8d-0a20eab63bcc\") " Dec 10 19:30:03 crc kubenswrapper[4828]: I1210 19:30:03.104350 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5hms\" (UniqueName: \"kubernetes.io/projected/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-kube-api-access-q5hms\") pod \"e8499d18-c5eb-4373-ab8d-0a20eab63bcc\" (UID: \"e8499d18-c5eb-4373-ab8d-0a20eab63bcc\") " Dec 10 19:30:03 crc kubenswrapper[4828]: I1210 19:30:03.104461 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-config-volume" (OuterVolumeSpecName: "config-volume") pod "e8499d18-c5eb-4373-ab8d-0a20eab63bcc" (UID: "e8499d18-c5eb-4373-ab8d-0a20eab63bcc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:30:03 crc kubenswrapper[4828]: I1210 19:30:03.105118 4828 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:03 crc kubenswrapper[4828]: I1210 19:30:03.110708 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-kube-api-access-q5hms" (OuterVolumeSpecName: "kube-api-access-q5hms") pod "e8499d18-c5eb-4373-ab8d-0a20eab63bcc" (UID: "e8499d18-c5eb-4373-ab8d-0a20eab63bcc"). InnerVolumeSpecName "kube-api-access-q5hms". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:30:03 crc kubenswrapper[4828]: I1210 19:30:03.111684 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e8499d18-c5eb-4373-ab8d-0a20eab63bcc" (UID: "e8499d18-c5eb-4373-ab8d-0a20eab63bcc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:30:03 crc kubenswrapper[4828]: I1210 19:30:03.210599 4828 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:03 crc kubenswrapper[4828]: I1210 19:30:03.210848 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5hms\" (UniqueName: \"kubernetes.io/projected/e8499d18-c5eb-4373-ab8d-0a20eab63bcc-kube-api-access-q5hms\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:03 crc kubenswrapper[4828]: I1210 19:30:03.594572 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj" event={"ID":"e8499d18-c5eb-4373-ab8d-0a20eab63bcc","Type":"ContainerDied","Data":"bc461209f1b3ccae625611bd36d2b02a5f2d2d830604164a13c656424572a999"} Dec 10 19:30:03 crc kubenswrapper[4828]: I1210 19:30:03.594619 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc461209f1b3ccae625611bd36d2b02a5f2d2d830604164a13c656424572a999" Dec 10 19:30:03 crc kubenswrapper[4828]: I1210 19:30:03.594699 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj" Dec 10 19:30:04 crc kubenswrapper[4828]: I1210 19:30:04.044016 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj"] Dec 10 19:30:04 crc kubenswrapper[4828]: I1210 19:30:04.058453 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423205-s2lmj"] Dec 10 19:30:05 crc kubenswrapper[4828]: I1210 19:30:05.804545 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20d80266-84dc-4dee-83b6-3ccd9148fbca" path="/var/lib/kubelet/pods/20d80266-84dc-4dee-83b6-3ccd9148fbca/volumes" Dec 10 19:30:10 crc kubenswrapper[4828]: I1210 19:30:10.029237 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-2s89p"] Dec 10 19:30:10 crc kubenswrapper[4828]: I1210 19:30:10.040130 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-2s89p"] Dec 10 19:30:11 crc kubenswrapper[4828]: I1210 19:30:11.035002 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-gwbxv"] Dec 10 19:30:11 crc kubenswrapper[4828]: I1210 19:30:11.047775 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-gwbxv"] Dec 10 19:30:11 crc kubenswrapper[4828]: I1210 19:30:11.806932 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a634b8e-6ee2-4830-a764-2fcc5300b15d" path="/var/lib/kubelet/pods/1a634b8e-6ee2-4830-a764-2fcc5300b15d/volumes" Dec 10 19:30:11 crc kubenswrapper[4828]: I1210 19:30:11.807728 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a" path="/var/lib/kubelet/pods/1f2eb8fc-198b-4cf5-bcc9-e47f35b6324a/volumes" Dec 10 19:30:12 crc kubenswrapper[4828]: I1210 19:30:12.042127 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-n7jdh"] Dec 10 19:30:12 crc kubenswrapper[4828]: I1210 19:30:12.064004 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-n7jdh"] Dec 10 19:30:13 crc kubenswrapper[4828]: I1210 19:30:13.801484 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a56b6762-0c16-4b35-a11b-16e34810aeab" path="/var/lib/kubelet/pods/a56b6762-0c16-4b35-a11b-16e34810aeab/volumes" Dec 10 19:30:35 crc kubenswrapper[4828]: I1210 19:30:35.041784 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-nkxxv"] Dec 10 19:30:35 crc kubenswrapper[4828]: I1210 19:30:35.053067 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-nkxxv"] Dec 10 19:30:35 crc kubenswrapper[4828]: I1210 19:30:35.802332 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b" path="/var/lib/kubelet/pods/d9f4b1e5-d6ed-41d0-ac4e-22f4ff211a8b/volumes" Dec 10 19:30:38 crc kubenswrapper[4828]: I1210 19:30:38.027746 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-sqwjh"] Dec 10 19:30:38 crc kubenswrapper[4828]: I1210 19:30:38.038596 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-sqwjh"] Dec 10 19:30:39 crc kubenswrapper[4828]: I1210 19:30:39.803041 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22554fda-f55d-4243-8416-7362ae946396" path="/var/lib/kubelet/pods/22554fda-f55d-4243-8416-7362ae946396/volumes" Dec 10 19:30:40 crc kubenswrapper[4828]: I1210 19:30:40.402070 4828 scope.go:117] "RemoveContainer" containerID="fb43fc1f48ae34006bf74d4d8397051d5eeda58f610595c35d22768fdbb8d948" Dec 10 19:30:40 crc kubenswrapper[4828]: I1210 19:30:40.432703 4828 scope.go:117] "RemoveContainer" containerID="da2f7b7fd47c67b352b34c029147284bbad04762c536da585a6f875dbf0617d8" Dec 10 19:30:40 crc kubenswrapper[4828]: I1210 19:30:40.498499 4828 scope.go:117] "RemoveContainer" containerID="fc6c64ed594fe07d98e5a50b61dc2fe5a667292c054740f5de9c8d1fe0244a74" Dec 10 19:30:40 crc kubenswrapper[4828]: I1210 19:30:40.564905 4828 scope.go:117] "RemoveContainer" containerID="c5f3ce88ff6d90ede4552698a8c5ec775f1d143ab5228560c41b371330fc2ff1" Dec 10 19:30:40 crc kubenswrapper[4828]: I1210 19:30:40.604925 4828 scope.go:117] "RemoveContainer" containerID="19ee6e83dd0796c96ccc79d946b07dfd0afa1e4bbe6e91eecc44d95e8779a979" Dec 10 19:30:40 crc kubenswrapper[4828]: I1210 19:30:40.686276 4828 scope.go:117] "RemoveContainer" containerID="19fc8ad8f22009d9d47bef05d78283f3ad4397395f0f7469f014515fd84ccd08" Dec 10 19:30:53 crc kubenswrapper[4828]: I1210 19:30:53.101937 4828 generic.go:334] "Generic (PLEG): container finished" podID="98d9383c-0b06-4a8d-8d4c-1832493c3c5f" containerID="b41ce6891e3abf70029b38d2cb6e69a887ea8e687f32bbcb7e139c9ab73c7828" exitCode=0 Dec 10 19:30:53 crc kubenswrapper[4828]: I1210 19:30:53.102031 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" event={"ID":"98d9383c-0b06-4a8d-8d4c-1832493c3c5f","Type":"ContainerDied","Data":"b41ce6891e3abf70029b38d2cb6e69a887ea8e687f32bbcb7e139c9ab73c7828"} Dec 10 19:30:54 crc kubenswrapper[4828]: I1210 19:30:54.608169 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" Dec 10 19:30:54 crc kubenswrapper[4828]: I1210 19:30:54.725536 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-ssh-key\") pod \"98d9383c-0b06-4a8d-8d4c-1832493c3c5f\" (UID: \"98d9383c-0b06-4a8d-8d4c-1832493c3c5f\") " Dec 10 19:30:54 crc kubenswrapper[4828]: I1210 19:30:54.726244 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-inventory\") pod \"98d9383c-0b06-4a8d-8d4c-1832493c3c5f\" (UID: \"98d9383c-0b06-4a8d-8d4c-1832493c3c5f\") " Dec 10 19:30:54 crc kubenswrapper[4828]: I1210 19:30:54.726426 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mdvs\" (UniqueName: \"kubernetes.io/projected/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-kube-api-access-2mdvs\") pod \"98d9383c-0b06-4a8d-8d4c-1832493c3c5f\" (UID: \"98d9383c-0b06-4a8d-8d4c-1832493c3c5f\") " Dec 10 19:30:54 crc kubenswrapper[4828]: I1210 19:30:54.732011 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-kube-api-access-2mdvs" (OuterVolumeSpecName: "kube-api-access-2mdvs") pod "98d9383c-0b06-4a8d-8d4c-1832493c3c5f" (UID: "98d9383c-0b06-4a8d-8d4c-1832493c3c5f"). InnerVolumeSpecName "kube-api-access-2mdvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:30:54 crc kubenswrapper[4828]: I1210 19:30:54.757962 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-inventory" (OuterVolumeSpecName: "inventory") pod "98d9383c-0b06-4a8d-8d4c-1832493c3c5f" (UID: "98d9383c-0b06-4a8d-8d4c-1832493c3c5f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:30:54 crc kubenswrapper[4828]: I1210 19:30:54.759670 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "98d9383c-0b06-4a8d-8d4c-1832493c3c5f" (UID: "98d9383c-0b06-4a8d-8d4c-1832493c3c5f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:30:54 crc kubenswrapper[4828]: I1210 19:30:54.831079 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:54 crc kubenswrapper[4828]: I1210 19:30:54.831128 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:54 crc kubenswrapper[4828]: I1210 19:30:54.831144 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mdvs\" (UniqueName: \"kubernetes.io/projected/98d9383c-0b06-4a8d-8d4c-1832493c3c5f-kube-api-access-2mdvs\") on node \"crc\" DevicePath \"\"" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.122700 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" event={"ID":"98d9383c-0b06-4a8d-8d4c-1832493c3c5f","Type":"ContainerDied","Data":"8f43073f50ac151571c7d4001354a83748b5d28556bc76d95d9ec1eda488ab7b"} Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.122747 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f43073f50ac151571c7d4001354a83748b5d28556bc76d95d9ec1eda488ab7b" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.122747 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.206528 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf"] Dec 10 19:30:55 crc kubenswrapper[4828]: E1210 19:30:55.207205 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8499d18-c5eb-4373-ab8d-0a20eab63bcc" containerName="collect-profiles" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.207229 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8499d18-c5eb-4373-ab8d-0a20eab63bcc" containerName="collect-profiles" Dec 10 19:30:55 crc kubenswrapper[4828]: E1210 19:30:55.207275 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98d9383c-0b06-4a8d-8d4c-1832493c3c5f" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.207285 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="98d9383c-0b06-4a8d-8d4c-1832493c3c5f" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.207547 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8499d18-c5eb-4373-ab8d-0a20eab63bcc" containerName="collect-profiles" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.207609 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="98d9383c-0b06-4a8d-8d4c-1832493c3c5f" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.208608 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.210459 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.211385 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.212048 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.218466 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf"] Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.219409 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.342394 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5klw\" (UniqueName: \"kubernetes.io/projected/88eb8a6c-2e65-42b9-bda9-89c45196fd75-kube-api-access-d5klw\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf\" (UID: \"88eb8a6c-2e65-42b9-bda9-89c45196fd75\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.342512 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88eb8a6c-2e65-42b9-bda9-89c45196fd75-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf\" (UID: \"88eb8a6c-2e65-42b9-bda9-89c45196fd75\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.342586 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88eb8a6c-2e65-42b9-bda9-89c45196fd75-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf\" (UID: \"88eb8a6c-2e65-42b9-bda9-89c45196fd75\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.444453 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88eb8a6c-2e65-42b9-bda9-89c45196fd75-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf\" (UID: \"88eb8a6c-2e65-42b9-bda9-89c45196fd75\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.444564 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88eb8a6c-2e65-42b9-bda9-89c45196fd75-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf\" (UID: \"88eb8a6c-2e65-42b9-bda9-89c45196fd75\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.444672 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5klw\" (UniqueName: \"kubernetes.io/projected/88eb8a6c-2e65-42b9-bda9-89c45196fd75-kube-api-access-d5klw\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf\" (UID: \"88eb8a6c-2e65-42b9-bda9-89c45196fd75\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.449859 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88eb8a6c-2e65-42b9-bda9-89c45196fd75-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf\" (UID: \"88eb8a6c-2e65-42b9-bda9-89c45196fd75\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.449858 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88eb8a6c-2e65-42b9-bda9-89c45196fd75-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf\" (UID: \"88eb8a6c-2e65-42b9-bda9-89c45196fd75\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.462209 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5klw\" (UniqueName: \"kubernetes.io/projected/88eb8a6c-2e65-42b9-bda9-89c45196fd75-kube-api-access-d5klw\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf\" (UID: \"88eb8a6c-2e65-42b9-bda9-89c45196fd75\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" Dec 10 19:30:55 crc kubenswrapper[4828]: I1210 19:30:55.534019 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" Dec 10 19:30:56 crc kubenswrapper[4828]: I1210 19:30:56.077175 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf"] Dec 10 19:30:56 crc kubenswrapper[4828]: I1210 19:30:56.132655 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" event={"ID":"88eb8a6c-2e65-42b9-bda9-89c45196fd75","Type":"ContainerStarted","Data":"b26dc6e1df2a11d5019d4d112d80fefcb315277c7501013e136f14ef0bd73ce6"} Dec 10 19:30:57 crc kubenswrapper[4828]: I1210 19:30:57.149573 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" event={"ID":"88eb8a6c-2e65-42b9-bda9-89c45196fd75","Type":"ContainerStarted","Data":"4b2618704b2e393c800db0dbc8c4727a1eb81e1db904ae3508f5d26037c9c1d0"} Dec 10 19:30:57 crc kubenswrapper[4828]: I1210 19:30:57.175741 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" podStartSLOduration=1.544727093 podStartE2EDuration="2.175719134s" podCreationTimestamp="2025-12-10 19:30:55 +0000 UTC" firstStartedPulling="2025-12-10 19:30:56.079377947 +0000 UTC m=+2136.589988952" lastFinishedPulling="2025-12-10 19:30:56.710369968 +0000 UTC m=+2137.220980993" observedRunningTime="2025-12-10 19:30:57.164087425 +0000 UTC m=+2137.674698430" watchObservedRunningTime="2025-12-10 19:30:57.175719134 +0000 UTC m=+2137.686330139" Dec 10 19:31:18 crc kubenswrapper[4828]: I1210 19:31:18.054933 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-jhdsp"] Dec 10 19:31:18 crc kubenswrapper[4828]: I1210 19:31:18.080756 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-jhdsp"] Dec 10 19:31:19 crc kubenswrapper[4828]: I1210 19:31:19.029414 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-e96e-account-create-update-8kssp"] Dec 10 19:31:19 crc kubenswrapper[4828]: I1210 19:31:19.040388 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-c67d-account-create-update-d6zs6"] Dec 10 19:31:19 crc kubenswrapper[4828]: I1210 19:31:19.053543 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-c67d-account-create-update-d6zs6"] Dec 10 19:31:19 crc kubenswrapper[4828]: I1210 19:31:19.063460 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-e96e-account-create-update-8kssp"] Dec 10 19:31:19 crc kubenswrapper[4828]: I1210 19:31:19.803234 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f5c519b-682a-430f-b1ed-ee93f6b7c9f2" path="/var/lib/kubelet/pods/1f5c519b-682a-430f-b1ed-ee93f6b7c9f2/volumes" Dec 10 19:31:19 crc kubenswrapper[4828]: I1210 19:31:19.804608 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47be5ead-d925-4058-a390-88426775ebfa" path="/var/lib/kubelet/pods/47be5ead-d925-4058-a390-88426775ebfa/volumes" Dec 10 19:31:19 crc kubenswrapper[4828]: I1210 19:31:19.805444 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df859d5f-d811-4be3-82d8-8327c7cfb851" path="/var/lib/kubelet/pods/df859d5f-d811-4be3-82d8-8327c7cfb851/volumes" Dec 10 19:31:20 crc kubenswrapper[4828]: I1210 19:31:20.039138 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-5tqrl"] Dec 10 19:31:20 crc kubenswrapper[4828]: I1210 19:31:20.051766 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-28cs6"] Dec 10 19:31:20 crc kubenswrapper[4828]: I1210 19:31:20.061668 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-edd8-account-create-update-f87t8"] Dec 10 19:31:20 crc kubenswrapper[4828]: I1210 19:31:20.071862 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-5tqrl"] Dec 10 19:31:20 crc kubenswrapper[4828]: I1210 19:31:20.083358 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-edd8-account-create-update-f87t8"] Dec 10 19:31:20 crc kubenswrapper[4828]: I1210 19:31:20.093626 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-28cs6"] Dec 10 19:31:21 crc kubenswrapper[4828]: I1210 19:31:21.230166 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:31:21 crc kubenswrapper[4828]: I1210 19:31:21.230479 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:31:21 crc kubenswrapper[4828]: I1210 19:31:21.802902 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b19ce251-54b7-40f3-bbe0-6afb6064b307" path="/var/lib/kubelet/pods/b19ce251-54b7-40f3-bbe0-6afb6064b307/volumes" Dec 10 19:31:21 crc kubenswrapper[4828]: I1210 19:31:21.803486 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b544108c-9253-40f1-a5ed-8456173a3e68" path="/var/lib/kubelet/pods/b544108c-9253-40f1-a5ed-8456173a3e68/volumes" Dec 10 19:31:21 crc kubenswrapper[4828]: I1210 19:31:21.804044 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d24aa875-f0bd-4e7d-a985-c633a246fb30" path="/var/lib/kubelet/pods/d24aa875-f0bd-4e7d-a985-c633a246fb30/volumes" Dec 10 19:31:40 crc kubenswrapper[4828]: I1210 19:31:40.835338 4828 scope.go:117] "RemoveContainer" containerID="bc65d15e3a6a8d0cd222585dce602e371e206cbf5ac9b84fb263d5b0433ef7d7" Dec 10 19:31:40 crc kubenswrapper[4828]: I1210 19:31:40.876343 4828 scope.go:117] "RemoveContainer" containerID="e6361c7b85657f7de6ac78e64a0b6a2bd350a593ed7056221d07e6c4b4c70376" Dec 10 19:31:40 crc kubenswrapper[4828]: I1210 19:31:40.920939 4828 scope.go:117] "RemoveContainer" containerID="1b6601f4598967d3b11414c3a4ffa6142b58840c11402f805f095360b0f6cc2e" Dec 10 19:31:40 crc kubenswrapper[4828]: I1210 19:31:40.980585 4828 scope.go:117] "RemoveContainer" containerID="cca446d297b8461f3e8fd45046e8da8e6e5e9193cadc282562fe2e510f32cc05" Dec 10 19:31:41 crc kubenswrapper[4828]: I1210 19:31:41.036341 4828 scope.go:117] "RemoveContainer" containerID="33d6df1cc9c23399a3e761f34b73e1aa89ae953a44e490b0c961b3bd25621b7f" Dec 10 19:31:41 crc kubenswrapper[4828]: I1210 19:31:41.087065 4828 scope.go:117] "RemoveContainer" containerID="2c3945b533871f06f437af5630054bead64d7ada9fe59b3de65d4ba734c03760" Dec 10 19:31:51 crc kubenswrapper[4828]: I1210 19:31:51.230285 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:31:51 crc kubenswrapper[4828]: I1210 19:31:51.230860 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:31:57 crc kubenswrapper[4828]: I1210 19:31:57.048692 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mzd68"] Dec 10 19:31:57 crc kubenswrapper[4828]: I1210 19:31:57.058452 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-mzd68"] Dec 10 19:31:57 crc kubenswrapper[4828]: I1210 19:31:57.802955 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff038e8c-07a0-41ca-9719-3b8e326721a0" path="/var/lib/kubelet/pods/ff038e8c-07a0-41ca-9719-3b8e326721a0/volumes" Dec 10 19:31:59 crc kubenswrapper[4828]: I1210 19:31:59.858794 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8zkvt"] Dec 10 19:31:59 crc kubenswrapper[4828]: I1210 19:31:59.862566 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zkvt" Dec 10 19:31:59 crc kubenswrapper[4828]: I1210 19:31:59.871701 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8zkvt"] Dec 10 19:32:00 crc kubenswrapper[4828]: I1210 19:32:00.020593 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f4f821-b7a1-496a-8268-066f62e33ad0-utilities\") pod \"redhat-operators-8zkvt\" (UID: \"34f4f821-b7a1-496a-8268-066f62e33ad0\") " pod="openshift-marketplace/redhat-operators-8zkvt" Dec 10 19:32:00 crc kubenswrapper[4828]: I1210 19:32:00.021054 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jxdw\" (UniqueName: \"kubernetes.io/projected/34f4f821-b7a1-496a-8268-066f62e33ad0-kube-api-access-2jxdw\") pod \"redhat-operators-8zkvt\" (UID: \"34f4f821-b7a1-496a-8268-066f62e33ad0\") " pod="openshift-marketplace/redhat-operators-8zkvt" Dec 10 19:32:00 crc kubenswrapper[4828]: I1210 19:32:00.021103 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f4f821-b7a1-496a-8268-066f62e33ad0-catalog-content\") pod \"redhat-operators-8zkvt\" (UID: \"34f4f821-b7a1-496a-8268-066f62e33ad0\") " pod="openshift-marketplace/redhat-operators-8zkvt" Dec 10 19:32:00 crc kubenswrapper[4828]: I1210 19:32:00.124188 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jxdw\" (UniqueName: \"kubernetes.io/projected/34f4f821-b7a1-496a-8268-066f62e33ad0-kube-api-access-2jxdw\") pod \"redhat-operators-8zkvt\" (UID: \"34f4f821-b7a1-496a-8268-066f62e33ad0\") " pod="openshift-marketplace/redhat-operators-8zkvt" Dec 10 19:32:00 crc kubenswrapper[4828]: I1210 19:32:00.124250 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f4f821-b7a1-496a-8268-066f62e33ad0-catalog-content\") pod \"redhat-operators-8zkvt\" (UID: \"34f4f821-b7a1-496a-8268-066f62e33ad0\") " pod="openshift-marketplace/redhat-operators-8zkvt" Dec 10 19:32:00 crc kubenswrapper[4828]: I1210 19:32:00.124382 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f4f821-b7a1-496a-8268-066f62e33ad0-utilities\") pod \"redhat-operators-8zkvt\" (UID: \"34f4f821-b7a1-496a-8268-066f62e33ad0\") " pod="openshift-marketplace/redhat-operators-8zkvt" Dec 10 19:32:00 crc kubenswrapper[4828]: I1210 19:32:00.124861 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f4f821-b7a1-496a-8268-066f62e33ad0-catalog-content\") pod \"redhat-operators-8zkvt\" (UID: \"34f4f821-b7a1-496a-8268-066f62e33ad0\") " pod="openshift-marketplace/redhat-operators-8zkvt" Dec 10 19:32:00 crc kubenswrapper[4828]: I1210 19:32:00.124998 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f4f821-b7a1-496a-8268-066f62e33ad0-utilities\") pod \"redhat-operators-8zkvt\" (UID: \"34f4f821-b7a1-496a-8268-066f62e33ad0\") " pod="openshift-marketplace/redhat-operators-8zkvt" Dec 10 19:32:00 crc kubenswrapper[4828]: I1210 19:32:00.146274 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jxdw\" (UniqueName: \"kubernetes.io/projected/34f4f821-b7a1-496a-8268-066f62e33ad0-kube-api-access-2jxdw\") pod \"redhat-operators-8zkvt\" (UID: \"34f4f821-b7a1-496a-8268-066f62e33ad0\") " pod="openshift-marketplace/redhat-operators-8zkvt" Dec 10 19:32:00 crc kubenswrapper[4828]: I1210 19:32:00.192882 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zkvt" Dec 10 19:32:00 crc kubenswrapper[4828]: I1210 19:32:00.670624 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8zkvt"] Dec 10 19:32:00 crc kubenswrapper[4828]: I1210 19:32:00.888443 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zkvt" event={"ID":"34f4f821-b7a1-496a-8268-066f62e33ad0","Type":"ContainerStarted","Data":"7ed7d5cef64f743a46bc591fae39812904531de80d05ef4b78e3bf7ffebb3a64"} Dec 10 19:32:01 crc kubenswrapper[4828]: I1210 19:32:01.900243 4828 generic.go:334] "Generic (PLEG): container finished" podID="34f4f821-b7a1-496a-8268-066f62e33ad0" containerID="50b112b71e85999bd417329e9562535f8a16179672cc26d8e3f82fb63a707e8e" exitCode=0 Dec 10 19:32:01 crc kubenswrapper[4828]: I1210 19:32:01.900309 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zkvt" event={"ID":"34f4f821-b7a1-496a-8268-066f62e33ad0","Type":"ContainerDied","Data":"50b112b71e85999bd417329e9562535f8a16179672cc26d8e3f82fb63a707e8e"} Dec 10 19:32:03 crc kubenswrapper[4828]: I1210 19:32:03.927318 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zkvt" event={"ID":"34f4f821-b7a1-496a-8268-066f62e33ad0","Type":"ContainerStarted","Data":"aa903e9ae3f1cbfa2bf291b96c13f3417d5fba5b42095ffdd6b620c5eb5ec678"} Dec 10 19:32:07 crc kubenswrapper[4828]: I1210 19:32:07.970535 4828 generic.go:334] "Generic (PLEG): container finished" podID="34f4f821-b7a1-496a-8268-066f62e33ad0" containerID="aa903e9ae3f1cbfa2bf291b96c13f3417d5fba5b42095ffdd6b620c5eb5ec678" exitCode=0 Dec 10 19:32:07 crc kubenswrapper[4828]: I1210 19:32:07.970613 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zkvt" event={"ID":"34f4f821-b7a1-496a-8268-066f62e33ad0","Type":"ContainerDied","Data":"aa903e9ae3f1cbfa2bf291b96c13f3417d5fba5b42095ffdd6b620c5eb5ec678"} Dec 10 19:32:09 crc kubenswrapper[4828]: I1210 19:32:09.996715 4828 generic.go:334] "Generic (PLEG): container finished" podID="88eb8a6c-2e65-42b9-bda9-89c45196fd75" containerID="4b2618704b2e393c800db0dbc8c4727a1eb81e1db904ae3508f5d26037c9c1d0" exitCode=0 Dec 10 19:32:09 crc kubenswrapper[4828]: I1210 19:32:09.996809 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" event={"ID":"88eb8a6c-2e65-42b9-bda9-89c45196fd75","Type":"ContainerDied","Data":"4b2618704b2e393c800db0dbc8c4727a1eb81e1db904ae3508f5d26037c9c1d0"} Dec 10 19:32:10 crc kubenswrapper[4828]: I1210 19:32:10.001265 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zkvt" event={"ID":"34f4f821-b7a1-496a-8268-066f62e33ad0","Type":"ContainerStarted","Data":"c7bbd413c97a2da29f5cbfa8ef3a72d6ca1eb04f270fef09ff31e52f9bb58e19"} Dec 10 19:32:10 crc kubenswrapper[4828]: I1210 19:32:10.045764 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8zkvt" podStartSLOduration=4.143618603 podStartE2EDuration="11.045744735s" podCreationTimestamp="2025-12-10 19:31:59 +0000 UTC" firstStartedPulling="2025-12-10 19:32:01.902210816 +0000 UTC m=+2202.412821821" lastFinishedPulling="2025-12-10 19:32:08.804336948 +0000 UTC m=+2209.314947953" observedRunningTime="2025-12-10 19:32:10.033610503 +0000 UTC m=+2210.544221508" watchObservedRunningTime="2025-12-10 19:32:10.045744735 +0000 UTC m=+2210.556355740" Dec 10 19:32:10 crc kubenswrapper[4828]: I1210 19:32:10.194109 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8zkvt" Dec 10 19:32:10 crc kubenswrapper[4828]: I1210 19:32:10.194159 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8zkvt" Dec 10 19:32:11 crc kubenswrapper[4828]: I1210 19:32:11.251064 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8zkvt" podUID="34f4f821-b7a1-496a-8268-066f62e33ad0" containerName="registry-server" probeResult="failure" output=< Dec 10 19:32:11 crc kubenswrapper[4828]: timeout: failed to connect service ":50051" within 1s Dec 10 19:32:11 crc kubenswrapper[4828]: > Dec 10 19:32:11 crc kubenswrapper[4828]: I1210 19:32:11.709884 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" Dec 10 19:32:11 crc kubenswrapper[4828]: I1210 19:32:11.809886 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5klw\" (UniqueName: \"kubernetes.io/projected/88eb8a6c-2e65-42b9-bda9-89c45196fd75-kube-api-access-d5klw\") pod \"88eb8a6c-2e65-42b9-bda9-89c45196fd75\" (UID: \"88eb8a6c-2e65-42b9-bda9-89c45196fd75\") " Dec 10 19:32:11 crc kubenswrapper[4828]: I1210 19:32:11.810284 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88eb8a6c-2e65-42b9-bda9-89c45196fd75-inventory\") pod \"88eb8a6c-2e65-42b9-bda9-89c45196fd75\" (UID: \"88eb8a6c-2e65-42b9-bda9-89c45196fd75\") " Dec 10 19:32:11 crc kubenswrapper[4828]: I1210 19:32:11.810324 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88eb8a6c-2e65-42b9-bda9-89c45196fd75-ssh-key\") pod \"88eb8a6c-2e65-42b9-bda9-89c45196fd75\" (UID: \"88eb8a6c-2e65-42b9-bda9-89c45196fd75\") " Dec 10 19:32:11 crc kubenswrapper[4828]: I1210 19:32:11.816307 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88eb8a6c-2e65-42b9-bda9-89c45196fd75-kube-api-access-d5klw" (OuterVolumeSpecName: "kube-api-access-d5klw") pod "88eb8a6c-2e65-42b9-bda9-89c45196fd75" (UID: "88eb8a6c-2e65-42b9-bda9-89c45196fd75"). InnerVolumeSpecName "kube-api-access-d5klw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:32:11 crc kubenswrapper[4828]: I1210 19:32:11.848509 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88eb8a6c-2e65-42b9-bda9-89c45196fd75-inventory" (OuterVolumeSpecName: "inventory") pod "88eb8a6c-2e65-42b9-bda9-89c45196fd75" (UID: "88eb8a6c-2e65-42b9-bda9-89c45196fd75"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:32:11 crc kubenswrapper[4828]: I1210 19:32:11.854715 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88eb8a6c-2e65-42b9-bda9-89c45196fd75-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "88eb8a6c-2e65-42b9-bda9-89c45196fd75" (UID: "88eb8a6c-2e65-42b9-bda9-89c45196fd75"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:32:11 crc kubenswrapper[4828]: I1210 19:32:11.912895 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5klw\" (UniqueName: \"kubernetes.io/projected/88eb8a6c-2e65-42b9-bda9-89c45196fd75-kube-api-access-d5klw\") on node \"crc\" DevicePath \"\"" Dec 10 19:32:11 crc kubenswrapper[4828]: I1210 19:32:11.912935 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/88eb8a6c-2e65-42b9-bda9-89c45196fd75-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:32:11 crc kubenswrapper[4828]: I1210 19:32:11.912944 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/88eb8a6c-2e65-42b9-bda9-89c45196fd75-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.024273 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" event={"ID":"88eb8a6c-2e65-42b9-bda9-89c45196fd75","Type":"ContainerDied","Data":"b26dc6e1df2a11d5019d4d112d80fefcb315277c7501013e136f14ef0bd73ce6"} Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.024319 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b26dc6e1df2a11d5019d4d112d80fefcb315277c7501013e136f14ef0bd73ce6" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.024394 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.213384 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt"] Dec 10 19:32:12 crc kubenswrapper[4828]: E1210 19:32:12.213906 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88eb8a6c-2e65-42b9-bda9-89c45196fd75" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.213926 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="88eb8a6c-2e65-42b9-bda9-89c45196fd75" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.214172 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="88eb8a6c-2e65-42b9-bda9-89c45196fd75" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.215068 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.221215 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.221329 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.221357 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.221400 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.230708 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt"] Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.321724 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e0b0f30-65f8-48bf-9dbf-223196457d89-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt\" (UID: \"0e0b0f30-65f8-48bf-9dbf-223196457d89\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.321779 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e0b0f30-65f8-48bf-9dbf-223196457d89-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt\" (UID: \"0e0b0f30-65f8-48bf-9dbf-223196457d89\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.322001 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8pjp\" (UniqueName: \"kubernetes.io/projected/0e0b0f30-65f8-48bf-9dbf-223196457d89-kube-api-access-q8pjp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt\" (UID: \"0e0b0f30-65f8-48bf-9dbf-223196457d89\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.424161 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8pjp\" (UniqueName: \"kubernetes.io/projected/0e0b0f30-65f8-48bf-9dbf-223196457d89-kube-api-access-q8pjp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt\" (UID: \"0e0b0f30-65f8-48bf-9dbf-223196457d89\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.424257 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e0b0f30-65f8-48bf-9dbf-223196457d89-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt\" (UID: \"0e0b0f30-65f8-48bf-9dbf-223196457d89\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.424290 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e0b0f30-65f8-48bf-9dbf-223196457d89-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt\" (UID: \"0e0b0f30-65f8-48bf-9dbf-223196457d89\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.437418 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e0b0f30-65f8-48bf-9dbf-223196457d89-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt\" (UID: \"0e0b0f30-65f8-48bf-9dbf-223196457d89\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.437418 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e0b0f30-65f8-48bf-9dbf-223196457d89-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt\" (UID: \"0e0b0f30-65f8-48bf-9dbf-223196457d89\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.440957 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8pjp\" (UniqueName: \"kubernetes.io/projected/0e0b0f30-65f8-48bf-9dbf-223196457d89-kube-api-access-q8pjp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt\" (UID: \"0e0b0f30-65f8-48bf-9dbf-223196457d89\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" Dec 10 19:32:12 crc kubenswrapper[4828]: I1210 19:32:12.536704 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" Dec 10 19:32:13 crc kubenswrapper[4828]: W1210 19:32:13.164005 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e0b0f30_65f8_48bf_9dbf_223196457d89.slice/crio-332a1fc1f48064ca100d8a633ca3cc8e1bbfe4b60e91c40f6d603a452d141baf WatchSource:0}: Error finding container 332a1fc1f48064ca100d8a633ca3cc8e1bbfe4b60e91c40f6d603a452d141baf: Status 404 returned error can't find the container with id 332a1fc1f48064ca100d8a633ca3cc8e1bbfe4b60e91c40f6d603a452d141baf Dec 10 19:32:13 crc kubenswrapper[4828]: I1210 19:32:13.165180 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt"] Dec 10 19:32:14 crc kubenswrapper[4828]: I1210 19:32:14.040105 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-j8658"] Dec 10 19:32:14 crc kubenswrapper[4828]: I1210 19:32:14.055141 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" event={"ID":"0e0b0f30-65f8-48bf-9dbf-223196457d89","Type":"ContainerStarted","Data":"f675f4462bcebfb4e5bf245f9104db546f393ec321f1518bb04a27d34bcd2be4"} Dec 10 19:32:14 crc kubenswrapper[4828]: I1210 19:32:14.055185 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" event={"ID":"0e0b0f30-65f8-48bf-9dbf-223196457d89","Type":"ContainerStarted","Data":"332a1fc1f48064ca100d8a633ca3cc8e1bbfe4b60e91c40f6d603a452d141baf"} Dec 10 19:32:14 crc kubenswrapper[4828]: I1210 19:32:14.057311 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-be42-account-create-update-j8bfv"] Dec 10 19:32:14 crc kubenswrapper[4828]: I1210 19:32:14.067266 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-j8658"] Dec 10 19:32:14 crc kubenswrapper[4828]: I1210 19:32:14.079884 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-be42-account-create-update-j8bfv"] Dec 10 19:32:14 crc kubenswrapper[4828]: I1210 19:32:14.082908 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" podStartSLOduration=1.54299399 podStartE2EDuration="2.082888695s" podCreationTimestamp="2025-12-10 19:32:12 +0000 UTC" firstStartedPulling="2025-12-10 19:32:13.167526189 +0000 UTC m=+2213.678137194" lastFinishedPulling="2025-12-10 19:32:13.707420884 +0000 UTC m=+2214.218031899" observedRunningTime="2025-12-10 19:32:14.072188951 +0000 UTC m=+2214.582799966" watchObservedRunningTime="2025-12-10 19:32:14.082888695 +0000 UTC m=+2214.593499700" Dec 10 19:32:15 crc kubenswrapper[4828]: I1210 19:32:15.801963 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9a81ca2-b1ab-4f73-a928-332933d6dd23" path="/var/lib/kubelet/pods/b9a81ca2-b1ab-4f73-a928-332933d6dd23/volumes" Dec 10 19:32:15 crc kubenswrapper[4828]: I1210 19:32:15.803271 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf3c61d9-4edd-4fe5-a591-7b3ef8455b13" path="/var/lib/kubelet/pods/cf3c61d9-4edd-4fe5-a591-7b3ef8455b13/volumes" Dec 10 19:32:19 crc kubenswrapper[4828]: I1210 19:32:19.107413 4828 generic.go:334] "Generic (PLEG): container finished" podID="0e0b0f30-65f8-48bf-9dbf-223196457d89" containerID="f675f4462bcebfb4e5bf245f9104db546f393ec321f1518bb04a27d34bcd2be4" exitCode=0 Dec 10 19:32:19 crc kubenswrapper[4828]: I1210 19:32:19.107487 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" event={"ID":"0e0b0f30-65f8-48bf-9dbf-223196457d89","Type":"ContainerDied","Data":"f675f4462bcebfb4e5bf245f9104db546f393ec321f1518bb04a27d34bcd2be4"} Dec 10 19:32:20 crc kubenswrapper[4828]: I1210 19:32:20.587560 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" Dec 10 19:32:20 crc kubenswrapper[4828]: I1210 19:32:20.747226 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8pjp\" (UniqueName: \"kubernetes.io/projected/0e0b0f30-65f8-48bf-9dbf-223196457d89-kube-api-access-q8pjp\") pod \"0e0b0f30-65f8-48bf-9dbf-223196457d89\" (UID: \"0e0b0f30-65f8-48bf-9dbf-223196457d89\") " Dec 10 19:32:20 crc kubenswrapper[4828]: I1210 19:32:20.747540 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e0b0f30-65f8-48bf-9dbf-223196457d89-inventory\") pod \"0e0b0f30-65f8-48bf-9dbf-223196457d89\" (UID: \"0e0b0f30-65f8-48bf-9dbf-223196457d89\") " Dec 10 19:32:20 crc kubenswrapper[4828]: I1210 19:32:20.747687 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e0b0f30-65f8-48bf-9dbf-223196457d89-ssh-key\") pod \"0e0b0f30-65f8-48bf-9dbf-223196457d89\" (UID: \"0e0b0f30-65f8-48bf-9dbf-223196457d89\") " Dec 10 19:32:20 crc kubenswrapper[4828]: I1210 19:32:20.753885 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e0b0f30-65f8-48bf-9dbf-223196457d89-kube-api-access-q8pjp" (OuterVolumeSpecName: "kube-api-access-q8pjp") pod "0e0b0f30-65f8-48bf-9dbf-223196457d89" (UID: "0e0b0f30-65f8-48bf-9dbf-223196457d89"). InnerVolumeSpecName "kube-api-access-q8pjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:32:20 crc kubenswrapper[4828]: I1210 19:32:20.779476 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e0b0f30-65f8-48bf-9dbf-223196457d89-inventory" (OuterVolumeSpecName: "inventory") pod "0e0b0f30-65f8-48bf-9dbf-223196457d89" (UID: "0e0b0f30-65f8-48bf-9dbf-223196457d89"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:32:20 crc kubenswrapper[4828]: I1210 19:32:20.779705 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e0b0f30-65f8-48bf-9dbf-223196457d89-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0e0b0f30-65f8-48bf-9dbf-223196457d89" (UID: "0e0b0f30-65f8-48bf-9dbf-223196457d89"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:32:20 crc kubenswrapper[4828]: I1210 19:32:20.850226 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8pjp\" (UniqueName: \"kubernetes.io/projected/0e0b0f30-65f8-48bf-9dbf-223196457d89-kube-api-access-q8pjp\") on node \"crc\" DevicePath \"\"" Dec 10 19:32:20 crc kubenswrapper[4828]: I1210 19:32:20.850264 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e0b0f30-65f8-48bf-9dbf-223196457d89-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:32:20 crc kubenswrapper[4828]: I1210 19:32:20.850274 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e0b0f30-65f8-48bf-9dbf-223196457d89-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.136771 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" event={"ID":"0e0b0f30-65f8-48bf-9dbf-223196457d89","Type":"ContainerDied","Data":"332a1fc1f48064ca100d8a633ca3cc8e1bbfe4b60e91c40f6d603a452d141baf"} Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.136856 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="332a1fc1f48064ca100d8a633ca3cc8e1bbfe4b60e91c40f6d603a452d141baf" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.136823 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.200951 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm"] Dec 10 19:32:21 crc kubenswrapper[4828]: E1210 19:32:21.201736 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e0b0f30-65f8-48bf-9dbf-223196457d89" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.201756 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e0b0f30-65f8-48bf-9dbf-223196457d89" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.202128 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e0b0f30-65f8-48bf-9dbf-223196457d89" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.202916 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.206342 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.206497 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.206930 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.207204 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.221679 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm"] Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.230565 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.230606 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.230643 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.231475 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.231531 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" gracePeriod=600 Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.252592 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8zkvt" podUID="34f4f821-b7a1-496a-8268-066f62e33ad0" containerName="registry-server" probeResult="failure" output=< Dec 10 19:32:21 crc kubenswrapper[4828]: timeout: failed to connect service ":50051" within 1s Dec 10 19:32:21 crc kubenswrapper[4828]: > Dec 10 19:32:21 crc kubenswrapper[4828]: E1210 19:32:21.360860 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.363419 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-snmlm\" (UID: \"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.363574 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnmqz\" (UniqueName: \"kubernetes.io/projected/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-kube-api-access-xnmqz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-snmlm\" (UID: \"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.364774 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-snmlm\" (UID: \"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.467447 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-snmlm\" (UID: \"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.467637 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-snmlm\" (UID: \"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.467689 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnmqz\" (UniqueName: \"kubernetes.io/projected/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-kube-api-access-xnmqz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-snmlm\" (UID: \"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.471842 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-snmlm\" (UID: \"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.472394 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-snmlm\" (UID: \"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.493584 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnmqz\" (UniqueName: \"kubernetes.io/projected/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-kube-api-access-xnmqz\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-snmlm\" (UID: \"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" Dec 10 19:32:21 crc kubenswrapper[4828]: I1210 19:32:21.522442 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" Dec 10 19:32:22 crc kubenswrapper[4828]: I1210 19:32:22.041940 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm"] Dec 10 19:32:22 crc kubenswrapper[4828]: W1210 19:32:22.044188 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2bb52b6f_bc4a_411d_be87_eb45eccbeb4a.slice/crio-6162cd86fb7698c0d2f207f1ac1a60cd90ddf000a501023ed6de2ed82281f86d WatchSource:0}: Error finding container 6162cd86fb7698c0d2f207f1ac1a60cd90ddf000a501023ed6de2ed82281f86d: Status 404 returned error can't find the container with id 6162cd86fb7698c0d2f207f1ac1a60cd90ddf000a501023ed6de2ed82281f86d Dec 10 19:32:22 crc kubenswrapper[4828]: I1210 19:32:22.148882 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" exitCode=0 Dec 10 19:32:22 crc kubenswrapper[4828]: I1210 19:32:22.148919 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6"} Dec 10 19:32:22 crc kubenswrapper[4828]: I1210 19:32:22.148990 4828 scope.go:117] "RemoveContainer" containerID="b4415f07501c6dc89d7e4fe4d4e51bb63a0dedd7e1962e15bc82fe0414c7a987" Dec 10 19:32:22 crc kubenswrapper[4828]: I1210 19:32:22.150575 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:32:22 crc kubenswrapper[4828]: E1210 19:32:22.150883 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:32:22 crc kubenswrapper[4828]: I1210 19:32:22.151263 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" event={"ID":"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a","Type":"ContainerStarted","Data":"6162cd86fb7698c0d2f207f1ac1a60cd90ddf000a501023ed6de2ed82281f86d"} Dec 10 19:32:24 crc kubenswrapper[4828]: I1210 19:32:24.177376 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" event={"ID":"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a","Type":"ContainerStarted","Data":"d6dc61ce52dfc0cf9af998d483239ddf83831f14898b5cc06a28fa9c1f59e327"} Dec 10 19:32:24 crc kubenswrapper[4828]: I1210 19:32:24.197387 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" podStartSLOduration=1.630089564 podStartE2EDuration="3.197369815s" podCreationTimestamp="2025-12-10 19:32:21 +0000 UTC" firstStartedPulling="2025-12-10 19:32:22.046911771 +0000 UTC m=+2222.557522776" lastFinishedPulling="2025-12-10 19:32:23.614192002 +0000 UTC m=+2224.124803027" observedRunningTime="2025-12-10 19:32:24.196245105 +0000 UTC m=+2224.706856110" watchObservedRunningTime="2025-12-10 19:32:24.197369815 +0000 UTC m=+2224.707980820" Dec 10 19:32:27 crc kubenswrapper[4828]: I1210 19:32:27.043883 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-qkkq9"] Dec 10 19:32:27 crc kubenswrapper[4828]: I1210 19:32:27.054516 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-qkkq9"] Dec 10 19:32:27 crc kubenswrapper[4828]: I1210 19:32:27.803502 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5e369bc-b3a1-444c-a16d-e70c0b754cef" path="/var/lib/kubelet/pods/a5e369bc-b3a1-444c-a16d-e70c0b754cef/volumes" Dec 10 19:32:28 crc kubenswrapper[4828]: I1210 19:32:28.031041 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wr6ll"] Dec 10 19:32:28 crc kubenswrapper[4828]: I1210 19:32:28.045187 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wr6ll"] Dec 10 19:32:29 crc kubenswrapper[4828]: I1210 19:32:29.803761 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0793df8-d37f-4645-8f1d-3cc501d1ce40" path="/var/lib/kubelet/pods/f0793df8-d37f-4645-8f1d-3cc501d1ce40/volumes" Dec 10 19:32:30 crc kubenswrapper[4828]: I1210 19:32:30.282491 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8zkvt" Dec 10 19:32:30 crc kubenswrapper[4828]: I1210 19:32:30.342326 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8zkvt" Dec 10 19:32:31 crc kubenswrapper[4828]: I1210 19:32:31.065058 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8zkvt"] Dec 10 19:32:32 crc kubenswrapper[4828]: I1210 19:32:32.261290 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8zkvt" podUID="34f4f821-b7a1-496a-8268-066f62e33ad0" containerName="registry-server" containerID="cri-o://c7bbd413c97a2da29f5cbfa8ef3a72d6ca1eb04f270fef09ff31e52f9bb58e19" gracePeriod=2 Dec 10 19:32:32 crc kubenswrapper[4828]: I1210 19:32:32.775647 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zkvt" Dec 10 19:32:32 crc kubenswrapper[4828]: I1210 19:32:32.931885 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f4f821-b7a1-496a-8268-066f62e33ad0-utilities\") pod \"34f4f821-b7a1-496a-8268-066f62e33ad0\" (UID: \"34f4f821-b7a1-496a-8268-066f62e33ad0\") " Dec 10 19:32:32 crc kubenswrapper[4828]: I1210 19:32:32.932181 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f4f821-b7a1-496a-8268-066f62e33ad0-catalog-content\") pod \"34f4f821-b7a1-496a-8268-066f62e33ad0\" (UID: \"34f4f821-b7a1-496a-8268-066f62e33ad0\") " Dec 10 19:32:32 crc kubenswrapper[4828]: I1210 19:32:32.932240 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jxdw\" (UniqueName: \"kubernetes.io/projected/34f4f821-b7a1-496a-8268-066f62e33ad0-kube-api-access-2jxdw\") pod \"34f4f821-b7a1-496a-8268-066f62e33ad0\" (UID: \"34f4f821-b7a1-496a-8268-066f62e33ad0\") " Dec 10 19:32:32 crc kubenswrapper[4828]: I1210 19:32:32.932611 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34f4f821-b7a1-496a-8268-066f62e33ad0-utilities" (OuterVolumeSpecName: "utilities") pod "34f4f821-b7a1-496a-8268-066f62e33ad0" (UID: "34f4f821-b7a1-496a-8268-066f62e33ad0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:32:32 crc kubenswrapper[4828]: I1210 19:32:32.933518 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f4f821-b7a1-496a-8268-066f62e33ad0-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:32:32 crc kubenswrapper[4828]: I1210 19:32:32.938259 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34f4f821-b7a1-496a-8268-066f62e33ad0-kube-api-access-2jxdw" (OuterVolumeSpecName: "kube-api-access-2jxdw") pod "34f4f821-b7a1-496a-8268-066f62e33ad0" (UID: "34f4f821-b7a1-496a-8268-066f62e33ad0"). InnerVolumeSpecName "kube-api-access-2jxdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.030116 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34f4f821-b7a1-496a-8268-066f62e33ad0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34f4f821-b7a1-496a-8268-066f62e33ad0" (UID: "34f4f821-b7a1-496a-8268-066f62e33ad0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.036269 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f4f821-b7a1-496a-8268-066f62e33ad0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.036307 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jxdw\" (UniqueName: \"kubernetes.io/projected/34f4f821-b7a1-496a-8268-066f62e33ad0-kube-api-access-2jxdw\") on node \"crc\" DevicePath \"\"" Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.274134 4828 generic.go:334] "Generic (PLEG): container finished" podID="34f4f821-b7a1-496a-8268-066f62e33ad0" containerID="c7bbd413c97a2da29f5cbfa8ef3a72d6ca1eb04f270fef09ff31e52f9bb58e19" exitCode=0 Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.274179 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zkvt" event={"ID":"34f4f821-b7a1-496a-8268-066f62e33ad0","Type":"ContainerDied","Data":"c7bbd413c97a2da29f5cbfa8ef3a72d6ca1eb04f270fef09ff31e52f9bb58e19"} Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.274213 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8zkvt" event={"ID":"34f4f821-b7a1-496a-8268-066f62e33ad0","Type":"ContainerDied","Data":"7ed7d5cef64f743a46bc591fae39812904531de80d05ef4b78e3bf7ffebb3a64"} Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.274212 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8zkvt" Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.274229 4828 scope.go:117] "RemoveContainer" containerID="c7bbd413c97a2da29f5cbfa8ef3a72d6ca1eb04f270fef09ff31e52f9bb58e19" Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.309248 4828 scope.go:117] "RemoveContainer" containerID="aa903e9ae3f1cbfa2bf291b96c13f3417d5fba5b42095ffdd6b620c5eb5ec678" Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.344197 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8zkvt"] Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.351705 4828 scope.go:117] "RemoveContainer" containerID="50b112b71e85999bd417329e9562535f8a16179672cc26d8e3f82fb63a707e8e" Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.357423 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8zkvt"] Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.396443 4828 scope.go:117] "RemoveContainer" containerID="c7bbd413c97a2da29f5cbfa8ef3a72d6ca1eb04f270fef09ff31e52f9bb58e19" Dec 10 19:32:33 crc kubenswrapper[4828]: E1210 19:32:33.397021 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7bbd413c97a2da29f5cbfa8ef3a72d6ca1eb04f270fef09ff31e52f9bb58e19\": container with ID starting with c7bbd413c97a2da29f5cbfa8ef3a72d6ca1eb04f270fef09ff31e52f9bb58e19 not found: ID does not exist" containerID="c7bbd413c97a2da29f5cbfa8ef3a72d6ca1eb04f270fef09ff31e52f9bb58e19" Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.397051 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7bbd413c97a2da29f5cbfa8ef3a72d6ca1eb04f270fef09ff31e52f9bb58e19"} err="failed to get container status \"c7bbd413c97a2da29f5cbfa8ef3a72d6ca1eb04f270fef09ff31e52f9bb58e19\": rpc error: code = NotFound desc = could not find container \"c7bbd413c97a2da29f5cbfa8ef3a72d6ca1eb04f270fef09ff31e52f9bb58e19\": container with ID starting with c7bbd413c97a2da29f5cbfa8ef3a72d6ca1eb04f270fef09ff31e52f9bb58e19 not found: ID does not exist" Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.397072 4828 scope.go:117] "RemoveContainer" containerID="aa903e9ae3f1cbfa2bf291b96c13f3417d5fba5b42095ffdd6b620c5eb5ec678" Dec 10 19:32:33 crc kubenswrapper[4828]: E1210 19:32:33.397519 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa903e9ae3f1cbfa2bf291b96c13f3417d5fba5b42095ffdd6b620c5eb5ec678\": container with ID starting with aa903e9ae3f1cbfa2bf291b96c13f3417d5fba5b42095ffdd6b620c5eb5ec678 not found: ID does not exist" containerID="aa903e9ae3f1cbfa2bf291b96c13f3417d5fba5b42095ffdd6b620c5eb5ec678" Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.397537 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa903e9ae3f1cbfa2bf291b96c13f3417d5fba5b42095ffdd6b620c5eb5ec678"} err="failed to get container status \"aa903e9ae3f1cbfa2bf291b96c13f3417d5fba5b42095ffdd6b620c5eb5ec678\": rpc error: code = NotFound desc = could not find container \"aa903e9ae3f1cbfa2bf291b96c13f3417d5fba5b42095ffdd6b620c5eb5ec678\": container with ID starting with aa903e9ae3f1cbfa2bf291b96c13f3417d5fba5b42095ffdd6b620c5eb5ec678 not found: ID does not exist" Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.397551 4828 scope.go:117] "RemoveContainer" containerID="50b112b71e85999bd417329e9562535f8a16179672cc26d8e3f82fb63a707e8e" Dec 10 19:32:33 crc kubenswrapper[4828]: E1210 19:32:33.402278 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50b112b71e85999bd417329e9562535f8a16179672cc26d8e3f82fb63a707e8e\": container with ID starting with 50b112b71e85999bd417329e9562535f8a16179672cc26d8e3f82fb63a707e8e not found: ID does not exist" containerID="50b112b71e85999bd417329e9562535f8a16179672cc26d8e3f82fb63a707e8e" Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.402329 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50b112b71e85999bd417329e9562535f8a16179672cc26d8e3f82fb63a707e8e"} err="failed to get container status \"50b112b71e85999bd417329e9562535f8a16179672cc26d8e3f82fb63a707e8e\": rpc error: code = NotFound desc = could not find container \"50b112b71e85999bd417329e9562535f8a16179672cc26d8e3f82fb63a707e8e\": container with ID starting with 50b112b71e85999bd417329e9562535f8a16179672cc26d8e3f82fb63a707e8e not found: ID does not exist" Dec 10 19:32:33 crc kubenswrapper[4828]: I1210 19:32:33.801338 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34f4f821-b7a1-496a-8268-066f62e33ad0" path="/var/lib/kubelet/pods/34f4f821-b7a1-496a-8268-066f62e33ad0/volumes" Dec 10 19:32:35 crc kubenswrapper[4828]: I1210 19:32:35.788727 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:32:35 crc kubenswrapper[4828]: E1210 19:32:35.789281 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:32:41 crc kubenswrapper[4828]: I1210 19:32:41.247962 4828 scope.go:117] "RemoveContainer" containerID="c0129558dfdd3e4177e968047ac2960485d6ab7fb4131ec5025c1a85d544ca18" Dec 10 19:32:41 crc kubenswrapper[4828]: I1210 19:32:41.280105 4828 scope.go:117] "RemoveContainer" containerID="878b0bf453660768bb6ad41edd6478b6244747f6bbeca165254f867dcf7a3a1f" Dec 10 19:32:41 crc kubenswrapper[4828]: I1210 19:32:41.347356 4828 scope.go:117] "RemoveContainer" containerID="5c4d6477387947edb0c421a01054cdf917c21c3a38f7a0d8a05bdcc45d8ef1b1" Dec 10 19:32:41 crc kubenswrapper[4828]: I1210 19:32:41.399227 4828 scope.go:117] "RemoveContainer" containerID="90281732151b276c44c07e2e044c7066937b465cc73bbef5ac0f509487deaeaf" Dec 10 19:32:41 crc kubenswrapper[4828]: I1210 19:32:41.468146 4828 scope.go:117] "RemoveContainer" containerID="4cba3f00e1c0289ab6e191fb3c7d9b20adfe56571354534de26f2d9cd52096a4" Dec 10 19:32:44 crc kubenswrapper[4828]: I1210 19:32:44.959489 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7tzjm"] Dec 10 19:32:44 crc kubenswrapper[4828]: E1210 19:32:44.960772 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34f4f821-b7a1-496a-8268-066f62e33ad0" containerName="extract-utilities" Dec 10 19:32:44 crc kubenswrapper[4828]: I1210 19:32:44.960791 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="34f4f821-b7a1-496a-8268-066f62e33ad0" containerName="extract-utilities" Dec 10 19:32:44 crc kubenswrapper[4828]: E1210 19:32:44.960834 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34f4f821-b7a1-496a-8268-066f62e33ad0" containerName="extract-content" Dec 10 19:32:44 crc kubenswrapper[4828]: I1210 19:32:44.960843 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="34f4f821-b7a1-496a-8268-066f62e33ad0" containerName="extract-content" Dec 10 19:32:44 crc kubenswrapper[4828]: E1210 19:32:44.960850 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34f4f821-b7a1-496a-8268-066f62e33ad0" containerName="registry-server" Dec 10 19:32:44 crc kubenswrapper[4828]: I1210 19:32:44.960858 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="34f4f821-b7a1-496a-8268-066f62e33ad0" containerName="registry-server" Dec 10 19:32:44 crc kubenswrapper[4828]: I1210 19:32:44.961137 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="34f4f821-b7a1-496a-8268-066f62e33ad0" containerName="registry-server" Dec 10 19:32:44 crc kubenswrapper[4828]: I1210 19:32:44.963000 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7tzjm" Dec 10 19:32:44 crc kubenswrapper[4828]: I1210 19:32:44.988958 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7tzjm"] Dec 10 19:32:45 crc kubenswrapper[4828]: I1210 19:32:45.017119 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ffcq\" (UniqueName: \"kubernetes.io/projected/30656d9d-2824-4cd7-81c0-aa5fc0697505-kube-api-access-6ffcq\") pod \"certified-operators-7tzjm\" (UID: \"30656d9d-2824-4cd7-81c0-aa5fc0697505\") " pod="openshift-marketplace/certified-operators-7tzjm" Dec 10 19:32:45 crc kubenswrapper[4828]: I1210 19:32:45.017397 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30656d9d-2824-4cd7-81c0-aa5fc0697505-utilities\") pod \"certified-operators-7tzjm\" (UID: \"30656d9d-2824-4cd7-81c0-aa5fc0697505\") " pod="openshift-marketplace/certified-operators-7tzjm" Dec 10 19:32:45 crc kubenswrapper[4828]: I1210 19:32:45.017421 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30656d9d-2824-4cd7-81c0-aa5fc0697505-catalog-content\") pod \"certified-operators-7tzjm\" (UID: \"30656d9d-2824-4cd7-81c0-aa5fc0697505\") " pod="openshift-marketplace/certified-operators-7tzjm" Dec 10 19:32:45 crc kubenswrapper[4828]: I1210 19:32:45.119951 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30656d9d-2824-4cd7-81c0-aa5fc0697505-utilities\") pod \"certified-operators-7tzjm\" (UID: \"30656d9d-2824-4cd7-81c0-aa5fc0697505\") " pod="openshift-marketplace/certified-operators-7tzjm" Dec 10 19:32:45 crc kubenswrapper[4828]: I1210 19:32:45.120004 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30656d9d-2824-4cd7-81c0-aa5fc0697505-catalog-content\") pod \"certified-operators-7tzjm\" (UID: \"30656d9d-2824-4cd7-81c0-aa5fc0697505\") " pod="openshift-marketplace/certified-operators-7tzjm" Dec 10 19:32:45 crc kubenswrapper[4828]: I1210 19:32:45.120095 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ffcq\" (UniqueName: \"kubernetes.io/projected/30656d9d-2824-4cd7-81c0-aa5fc0697505-kube-api-access-6ffcq\") pod \"certified-operators-7tzjm\" (UID: \"30656d9d-2824-4cd7-81c0-aa5fc0697505\") " pod="openshift-marketplace/certified-operators-7tzjm" Dec 10 19:32:45 crc kubenswrapper[4828]: I1210 19:32:45.120458 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30656d9d-2824-4cd7-81c0-aa5fc0697505-utilities\") pod \"certified-operators-7tzjm\" (UID: \"30656d9d-2824-4cd7-81c0-aa5fc0697505\") " pod="openshift-marketplace/certified-operators-7tzjm" Dec 10 19:32:45 crc kubenswrapper[4828]: I1210 19:32:45.120546 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30656d9d-2824-4cd7-81c0-aa5fc0697505-catalog-content\") pod \"certified-operators-7tzjm\" (UID: \"30656d9d-2824-4cd7-81c0-aa5fc0697505\") " pod="openshift-marketplace/certified-operators-7tzjm" Dec 10 19:32:45 crc kubenswrapper[4828]: I1210 19:32:45.144048 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ffcq\" (UniqueName: \"kubernetes.io/projected/30656d9d-2824-4cd7-81c0-aa5fc0697505-kube-api-access-6ffcq\") pod \"certified-operators-7tzjm\" (UID: \"30656d9d-2824-4cd7-81c0-aa5fc0697505\") " pod="openshift-marketplace/certified-operators-7tzjm" Dec 10 19:32:45 crc kubenswrapper[4828]: I1210 19:32:45.287940 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7tzjm" Dec 10 19:32:45 crc kubenswrapper[4828]: W1210 19:32:45.768530 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30656d9d_2824_4cd7_81c0_aa5fc0697505.slice/crio-bcd5feec234fade645d350d499255136d47643e143122d56f5a963cb47251693 WatchSource:0}: Error finding container bcd5feec234fade645d350d499255136d47643e143122d56f5a963cb47251693: Status 404 returned error can't find the container with id bcd5feec234fade645d350d499255136d47643e143122d56f5a963cb47251693 Dec 10 19:32:45 crc kubenswrapper[4828]: I1210 19:32:45.773639 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7tzjm"] Dec 10 19:32:46 crc kubenswrapper[4828]: I1210 19:32:46.502076 4828 generic.go:334] "Generic (PLEG): container finished" podID="30656d9d-2824-4cd7-81c0-aa5fc0697505" containerID="da6101d522337b6272eda6e836bd461d19e2286c9cf37e537ce4c15539d3e58f" exitCode=0 Dec 10 19:32:46 crc kubenswrapper[4828]: I1210 19:32:46.502187 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7tzjm" event={"ID":"30656d9d-2824-4cd7-81c0-aa5fc0697505","Type":"ContainerDied","Data":"da6101d522337b6272eda6e836bd461d19e2286c9cf37e537ce4c15539d3e58f"} Dec 10 19:32:46 crc kubenswrapper[4828]: I1210 19:32:46.502653 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7tzjm" event={"ID":"30656d9d-2824-4cd7-81c0-aa5fc0697505","Type":"ContainerStarted","Data":"bcd5feec234fade645d350d499255136d47643e143122d56f5a963cb47251693"} Dec 10 19:32:47 crc kubenswrapper[4828]: I1210 19:32:47.514411 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7tzjm" event={"ID":"30656d9d-2824-4cd7-81c0-aa5fc0697505","Type":"ContainerStarted","Data":"b1a1fa6a939142f3dcbfe9be0e0beaaedab65fc1ef7576ee384b84eef27ae61f"} Dec 10 19:32:48 crc kubenswrapper[4828]: I1210 19:32:48.788822 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:32:48 crc kubenswrapper[4828]: E1210 19:32:48.789332 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:32:49 crc kubenswrapper[4828]: I1210 19:32:49.031637 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-v9g52"] Dec 10 19:32:49 crc kubenswrapper[4828]: I1210 19:32:49.042118 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-v9g52"] Dec 10 19:32:49 crc kubenswrapper[4828]: I1210 19:32:49.535545 4828 generic.go:334] "Generic (PLEG): container finished" podID="30656d9d-2824-4cd7-81c0-aa5fc0697505" containerID="b1a1fa6a939142f3dcbfe9be0e0beaaedab65fc1ef7576ee384b84eef27ae61f" exitCode=0 Dec 10 19:32:49 crc kubenswrapper[4828]: I1210 19:32:49.535585 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7tzjm" event={"ID":"30656d9d-2824-4cd7-81c0-aa5fc0697505","Type":"ContainerDied","Data":"b1a1fa6a939142f3dcbfe9be0e0beaaedab65fc1ef7576ee384b84eef27ae61f"} Dec 10 19:32:49 crc kubenswrapper[4828]: I1210 19:32:49.803474 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac39304a-7ef4-42a7-a9b3-9ea15a237a4f" path="/var/lib/kubelet/pods/ac39304a-7ef4-42a7-a9b3-9ea15a237a4f/volumes" Dec 10 19:32:50 crc kubenswrapper[4828]: I1210 19:32:50.550178 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7tzjm" event={"ID":"30656d9d-2824-4cd7-81c0-aa5fc0697505","Type":"ContainerStarted","Data":"53085d2adefbd42538a66ae74b1021126182768c276bff4427da9c09c0b1e1d1"} Dec 10 19:32:55 crc kubenswrapper[4828]: I1210 19:32:55.288291 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7tzjm" Dec 10 19:32:55 crc kubenswrapper[4828]: I1210 19:32:55.289199 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7tzjm" Dec 10 19:32:55 crc kubenswrapper[4828]: I1210 19:32:55.346153 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7tzjm" Dec 10 19:32:55 crc kubenswrapper[4828]: I1210 19:32:55.375271 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7tzjm" podStartSLOduration=7.717048556 podStartE2EDuration="11.375251553s" podCreationTimestamp="2025-12-10 19:32:44 +0000 UTC" firstStartedPulling="2025-12-10 19:32:46.504291466 +0000 UTC m=+2247.014902471" lastFinishedPulling="2025-12-10 19:32:50.162494463 +0000 UTC m=+2250.673105468" observedRunningTime="2025-12-10 19:32:50.575288475 +0000 UTC m=+2251.085899520" watchObservedRunningTime="2025-12-10 19:32:55.375251553 +0000 UTC m=+2255.885862558" Dec 10 19:32:55 crc kubenswrapper[4828]: I1210 19:32:55.695216 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7tzjm" Dec 10 19:32:55 crc kubenswrapper[4828]: I1210 19:32:55.755987 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7tzjm"] Dec 10 19:32:57 crc kubenswrapper[4828]: I1210 19:32:57.632029 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7tzjm" podUID="30656d9d-2824-4cd7-81c0-aa5fc0697505" containerName="registry-server" containerID="cri-o://53085d2adefbd42538a66ae74b1021126182768c276bff4427da9c09c0b1e1d1" gracePeriod=2 Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.143409 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7tzjm" Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.239779 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30656d9d-2824-4cd7-81c0-aa5fc0697505-utilities\") pod \"30656d9d-2824-4cd7-81c0-aa5fc0697505\" (UID: \"30656d9d-2824-4cd7-81c0-aa5fc0697505\") " Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.240411 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ffcq\" (UniqueName: \"kubernetes.io/projected/30656d9d-2824-4cd7-81c0-aa5fc0697505-kube-api-access-6ffcq\") pod \"30656d9d-2824-4cd7-81c0-aa5fc0697505\" (UID: \"30656d9d-2824-4cd7-81c0-aa5fc0697505\") " Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.240485 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30656d9d-2824-4cd7-81c0-aa5fc0697505-catalog-content\") pod \"30656d9d-2824-4cd7-81c0-aa5fc0697505\" (UID: \"30656d9d-2824-4cd7-81c0-aa5fc0697505\") " Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.240677 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30656d9d-2824-4cd7-81c0-aa5fc0697505-utilities" (OuterVolumeSpecName: "utilities") pod "30656d9d-2824-4cd7-81c0-aa5fc0697505" (UID: "30656d9d-2824-4cd7-81c0-aa5fc0697505"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.240981 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30656d9d-2824-4cd7-81c0-aa5fc0697505-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.244825 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30656d9d-2824-4cd7-81c0-aa5fc0697505-kube-api-access-6ffcq" (OuterVolumeSpecName: "kube-api-access-6ffcq") pod "30656d9d-2824-4cd7-81c0-aa5fc0697505" (UID: "30656d9d-2824-4cd7-81c0-aa5fc0697505"). InnerVolumeSpecName "kube-api-access-6ffcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.288010 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30656d9d-2824-4cd7-81c0-aa5fc0697505-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30656d9d-2824-4cd7-81c0-aa5fc0697505" (UID: "30656d9d-2824-4cd7-81c0-aa5fc0697505"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.343616 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ffcq\" (UniqueName: \"kubernetes.io/projected/30656d9d-2824-4cd7-81c0-aa5fc0697505-kube-api-access-6ffcq\") on node \"crc\" DevicePath \"\"" Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.343648 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30656d9d-2824-4cd7-81c0-aa5fc0697505-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.647324 4828 generic.go:334] "Generic (PLEG): container finished" podID="30656d9d-2824-4cd7-81c0-aa5fc0697505" containerID="53085d2adefbd42538a66ae74b1021126182768c276bff4427da9c09c0b1e1d1" exitCode=0 Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.647368 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7tzjm" event={"ID":"30656d9d-2824-4cd7-81c0-aa5fc0697505","Type":"ContainerDied","Data":"53085d2adefbd42538a66ae74b1021126182768c276bff4427da9c09c0b1e1d1"} Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.647410 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7tzjm" event={"ID":"30656d9d-2824-4cd7-81c0-aa5fc0697505","Type":"ContainerDied","Data":"bcd5feec234fade645d350d499255136d47643e143122d56f5a963cb47251693"} Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.647428 4828 scope.go:117] "RemoveContainer" containerID="53085d2adefbd42538a66ae74b1021126182768c276bff4427da9c09c0b1e1d1" Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.647952 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7tzjm" Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.685506 4828 scope.go:117] "RemoveContainer" containerID="b1a1fa6a939142f3dcbfe9be0e0beaaedab65fc1ef7576ee384b84eef27ae61f" Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.698956 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7tzjm"] Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.708962 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7tzjm"] Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.714763 4828 scope.go:117] "RemoveContainer" containerID="da6101d522337b6272eda6e836bd461d19e2286c9cf37e537ce4c15539d3e58f" Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.782964 4828 scope.go:117] "RemoveContainer" containerID="53085d2adefbd42538a66ae74b1021126182768c276bff4427da9c09c0b1e1d1" Dec 10 19:32:58 crc kubenswrapper[4828]: E1210 19:32:58.783479 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53085d2adefbd42538a66ae74b1021126182768c276bff4427da9c09c0b1e1d1\": container with ID starting with 53085d2adefbd42538a66ae74b1021126182768c276bff4427da9c09c0b1e1d1 not found: ID does not exist" containerID="53085d2adefbd42538a66ae74b1021126182768c276bff4427da9c09c0b1e1d1" Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.783519 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53085d2adefbd42538a66ae74b1021126182768c276bff4427da9c09c0b1e1d1"} err="failed to get container status \"53085d2adefbd42538a66ae74b1021126182768c276bff4427da9c09c0b1e1d1\": rpc error: code = NotFound desc = could not find container \"53085d2adefbd42538a66ae74b1021126182768c276bff4427da9c09c0b1e1d1\": container with ID starting with 53085d2adefbd42538a66ae74b1021126182768c276bff4427da9c09c0b1e1d1 not found: ID does not exist" Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.783565 4828 scope.go:117] "RemoveContainer" containerID="b1a1fa6a939142f3dcbfe9be0e0beaaedab65fc1ef7576ee384b84eef27ae61f" Dec 10 19:32:58 crc kubenswrapper[4828]: E1210 19:32:58.784017 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1a1fa6a939142f3dcbfe9be0e0beaaedab65fc1ef7576ee384b84eef27ae61f\": container with ID starting with b1a1fa6a939142f3dcbfe9be0e0beaaedab65fc1ef7576ee384b84eef27ae61f not found: ID does not exist" containerID="b1a1fa6a939142f3dcbfe9be0e0beaaedab65fc1ef7576ee384b84eef27ae61f" Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.784067 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1a1fa6a939142f3dcbfe9be0e0beaaedab65fc1ef7576ee384b84eef27ae61f"} err="failed to get container status \"b1a1fa6a939142f3dcbfe9be0e0beaaedab65fc1ef7576ee384b84eef27ae61f\": rpc error: code = NotFound desc = could not find container \"b1a1fa6a939142f3dcbfe9be0e0beaaedab65fc1ef7576ee384b84eef27ae61f\": container with ID starting with b1a1fa6a939142f3dcbfe9be0e0beaaedab65fc1ef7576ee384b84eef27ae61f not found: ID does not exist" Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.784101 4828 scope.go:117] "RemoveContainer" containerID="da6101d522337b6272eda6e836bd461d19e2286c9cf37e537ce4c15539d3e58f" Dec 10 19:32:58 crc kubenswrapper[4828]: E1210 19:32:58.784357 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da6101d522337b6272eda6e836bd461d19e2286c9cf37e537ce4c15539d3e58f\": container with ID starting with da6101d522337b6272eda6e836bd461d19e2286c9cf37e537ce4c15539d3e58f not found: ID does not exist" containerID="da6101d522337b6272eda6e836bd461d19e2286c9cf37e537ce4c15539d3e58f" Dec 10 19:32:58 crc kubenswrapper[4828]: I1210 19:32:58.784420 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da6101d522337b6272eda6e836bd461d19e2286c9cf37e537ce4c15539d3e58f"} err="failed to get container status \"da6101d522337b6272eda6e836bd461d19e2286c9cf37e537ce4c15539d3e58f\": rpc error: code = NotFound desc = could not find container \"da6101d522337b6272eda6e836bd461d19e2286c9cf37e537ce4c15539d3e58f\": container with ID starting with da6101d522337b6272eda6e836bd461d19e2286c9cf37e537ce4c15539d3e58f not found: ID does not exist" Dec 10 19:32:59 crc kubenswrapper[4828]: I1210 19:32:59.802930 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30656d9d-2824-4cd7-81c0-aa5fc0697505" path="/var/lib/kubelet/pods/30656d9d-2824-4cd7-81c0-aa5fc0697505/volumes" Dec 10 19:33:00 crc kubenswrapper[4828]: I1210 19:33:00.790011 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:33:00 crc kubenswrapper[4828]: E1210 19:33:00.791030 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:33:02 crc kubenswrapper[4828]: I1210 19:33:02.720760 4828 generic.go:334] "Generic (PLEG): container finished" podID="2bb52b6f-bc4a-411d-be87-eb45eccbeb4a" containerID="d6dc61ce52dfc0cf9af998d483239ddf83831f14898b5cc06a28fa9c1f59e327" exitCode=0 Dec 10 19:33:02 crc kubenswrapper[4828]: I1210 19:33:02.721048 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" event={"ID":"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a","Type":"ContainerDied","Data":"d6dc61ce52dfc0cf9af998d483239ddf83831f14898b5cc06a28fa9c1f59e327"} Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.189928 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.376850 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnmqz\" (UniqueName: \"kubernetes.io/projected/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-kube-api-access-xnmqz\") pod \"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a\" (UID: \"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a\") " Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.376988 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-ssh-key\") pod \"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a\" (UID: \"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a\") " Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.377193 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-inventory\") pod \"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a\" (UID: \"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a\") " Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.382553 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-kube-api-access-xnmqz" (OuterVolumeSpecName: "kube-api-access-xnmqz") pod "2bb52b6f-bc4a-411d-be87-eb45eccbeb4a" (UID: "2bb52b6f-bc4a-411d-be87-eb45eccbeb4a"). InnerVolumeSpecName "kube-api-access-xnmqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.412038 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2bb52b6f-bc4a-411d-be87-eb45eccbeb4a" (UID: "2bb52b6f-bc4a-411d-be87-eb45eccbeb4a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.418833 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-inventory" (OuterVolumeSpecName: "inventory") pod "2bb52b6f-bc4a-411d-be87-eb45eccbeb4a" (UID: "2bb52b6f-bc4a-411d-be87-eb45eccbeb4a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.479687 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.479720 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnmqz\" (UniqueName: \"kubernetes.io/projected/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-kube-api-access-xnmqz\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.479730 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2bb52b6f-bc4a-411d-be87-eb45eccbeb4a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.743478 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" event={"ID":"2bb52b6f-bc4a-411d-be87-eb45eccbeb4a","Type":"ContainerDied","Data":"6162cd86fb7698c0d2f207f1ac1a60cd90ddf000a501023ed6de2ed82281f86d"} Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.743516 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6162cd86fb7698c0d2f207f1ac1a60cd90ddf000a501023ed6de2ed82281f86d" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.743576 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-snmlm" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.824501 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599"] Dec 10 19:33:04 crc kubenswrapper[4828]: E1210 19:33:04.825389 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30656d9d-2824-4cd7-81c0-aa5fc0697505" containerName="extract-utilities" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.825431 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="30656d9d-2824-4cd7-81c0-aa5fc0697505" containerName="extract-utilities" Dec 10 19:33:04 crc kubenswrapper[4828]: E1210 19:33:04.825448 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bb52b6f-bc4a-411d-be87-eb45eccbeb4a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.825457 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bb52b6f-bc4a-411d-be87-eb45eccbeb4a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:33:04 crc kubenswrapper[4828]: E1210 19:33:04.825481 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30656d9d-2824-4cd7-81c0-aa5fc0697505" containerName="extract-content" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.825489 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="30656d9d-2824-4cd7-81c0-aa5fc0697505" containerName="extract-content" Dec 10 19:33:04 crc kubenswrapper[4828]: E1210 19:33:04.825501 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30656d9d-2824-4cd7-81c0-aa5fc0697505" containerName="registry-server" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.825509 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="30656d9d-2824-4cd7-81c0-aa5fc0697505" containerName="registry-server" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.825793 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="30656d9d-2824-4cd7-81c0-aa5fc0697505" containerName="registry-server" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.825865 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bb52b6f-bc4a-411d-be87-eb45eccbeb4a" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.826869 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.829943 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.830035 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.830683 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.839831 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.848061 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599"] Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.887635 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3f83864-18ee-4493-b420-a536371c509b-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xd599\" (UID: \"b3f83864-18ee-4493-b420-a536371c509b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.887835 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2vkd\" (UniqueName: \"kubernetes.io/projected/b3f83864-18ee-4493-b420-a536371c509b-kube-api-access-j2vkd\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xd599\" (UID: \"b3f83864-18ee-4493-b420-a536371c509b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.888010 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3f83864-18ee-4493-b420-a536371c509b-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xd599\" (UID: \"b3f83864-18ee-4493-b420-a536371c509b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.990077 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3f83864-18ee-4493-b420-a536371c509b-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xd599\" (UID: \"b3f83864-18ee-4493-b420-a536371c509b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.990198 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2vkd\" (UniqueName: \"kubernetes.io/projected/b3f83864-18ee-4493-b420-a536371c509b-kube-api-access-j2vkd\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xd599\" (UID: \"b3f83864-18ee-4493-b420-a536371c509b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.990330 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3f83864-18ee-4493-b420-a536371c509b-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xd599\" (UID: \"b3f83864-18ee-4493-b420-a536371c509b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.994121 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3f83864-18ee-4493-b420-a536371c509b-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xd599\" (UID: \"b3f83864-18ee-4493-b420-a536371c509b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" Dec 10 19:33:04 crc kubenswrapper[4828]: I1210 19:33:04.994429 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3f83864-18ee-4493-b420-a536371c509b-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xd599\" (UID: \"b3f83864-18ee-4493-b420-a536371c509b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" Dec 10 19:33:05 crc kubenswrapper[4828]: I1210 19:33:05.005497 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2vkd\" (UniqueName: \"kubernetes.io/projected/b3f83864-18ee-4493-b420-a536371c509b-kube-api-access-j2vkd\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xd599\" (UID: \"b3f83864-18ee-4493-b420-a536371c509b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" Dec 10 19:33:05 crc kubenswrapper[4828]: I1210 19:33:05.145011 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" Dec 10 19:33:05 crc kubenswrapper[4828]: I1210 19:33:05.667099 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599"] Dec 10 19:33:05 crc kubenswrapper[4828]: W1210 19:33:05.669398 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3f83864_18ee_4493_b420_a536371c509b.slice/crio-35310d7018e2385a55a9ff474da62bc1b326dac247672d30e8058801a1c76547 WatchSource:0}: Error finding container 35310d7018e2385a55a9ff474da62bc1b326dac247672d30e8058801a1c76547: Status 404 returned error can't find the container with id 35310d7018e2385a55a9ff474da62bc1b326dac247672d30e8058801a1c76547 Dec 10 19:33:05 crc kubenswrapper[4828]: I1210 19:33:05.754752 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" event={"ID":"b3f83864-18ee-4493-b420-a536371c509b","Type":"ContainerStarted","Data":"35310d7018e2385a55a9ff474da62bc1b326dac247672d30e8058801a1c76547"} Dec 10 19:33:06 crc kubenswrapper[4828]: I1210 19:33:06.772489 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" event={"ID":"b3f83864-18ee-4493-b420-a536371c509b","Type":"ContainerStarted","Data":"244e4e284e09618028876df9e39983135fe5f1fdea0b0f4bf99441dbd21808b7"} Dec 10 19:33:06 crc kubenswrapper[4828]: I1210 19:33:06.791178 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" podStartSLOduration=2.321359896 podStartE2EDuration="2.79114952s" podCreationTimestamp="2025-12-10 19:33:04 +0000 UTC" firstStartedPulling="2025-12-10 19:33:05.672395038 +0000 UTC m=+2266.183006043" lastFinishedPulling="2025-12-10 19:33:06.142184632 +0000 UTC m=+2266.652795667" observedRunningTime="2025-12-10 19:33:06.787885604 +0000 UTC m=+2267.298496689" watchObservedRunningTime="2025-12-10 19:33:06.79114952 +0000 UTC m=+2267.301760555" Dec 10 19:33:11 crc kubenswrapper[4828]: I1210 19:33:11.788665 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:33:11 crc kubenswrapper[4828]: E1210 19:33:11.789445 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:33:21 crc kubenswrapper[4828]: I1210 19:33:21.462740 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zpcfk"] Dec 10 19:33:21 crc kubenswrapper[4828]: I1210 19:33:21.467957 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zpcfk" Dec 10 19:33:21 crc kubenswrapper[4828]: I1210 19:33:21.477074 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a58e124-b1be-43ef-a54d-8baa62a26014-utilities\") pod \"redhat-marketplace-zpcfk\" (UID: \"1a58e124-b1be-43ef-a54d-8baa62a26014\") " pod="openshift-marketplace/redhat-marketplace-zpcfk" Dec 10 19:33:21 crc kubenswrapper[4828]: I1210 19:33:21.477215 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4s5n\" (UniqueName: \"kubernetes.io/projected/1a58e124-b1be-43ef-a54d-8baa62a26014-kube-api-access-l4s5n\") pod \"redhat-marketplace-zpcfk\" (UID: \"1a58e124-b1be-43ef-a54d-8baa62a26014\") " pod="openshift-marketplace/redhat-marketplace-zpcfk" Dec 10 19:33:21 crc kubenswrapper[4828]: I1210 19:33:21.477648 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a58e124-b1be-43ef-a54d-8baa62a26014-catalog-content\") pod \"redhat-marketplace-zpcfk\" (UID: \"1a58e124-b1be-43ef-a54d-8baa62a26014\") " pod="openshift-marketplace/redhat-marketplace-zpcfk" Dec 10 19:33:21 crc kubenswrapper[4828]: I1210 19:33:21.481625 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zpcfk"] Dec 10 19:33:21 crc kubenswrapper[4828]: I1210 19:33:21.580966 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a58e124-b1be-43ef-a54d-8baa62a26014-catalog-content\") pod \"redhat-marketplace-zpcfk\" (UID: \"1a58e124-b1be-43ef-a54d-8baa62a26014\") " pod="openshift-marketplace/redhat-marketplace-zpcfk" Dec 10 19:33:21 crc kubenswrapper[4828]: I1210 19:33:21.581142 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a58e124-b1be-43ef-a54d-8baa62a26014-utilities\") pod \"redhat-marketplace-zpcfk\" (UID: \"1a58e124-b1be-43ef-a54d-8baa62a26014\") " pod="openshift-marketplace/redhat-marketplace-zpcfk" Dec 10 19:33:21 crc kubenswrapper[4828]: I1210 19:33:21.581198 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4s5n\" (UniqueName: \"kubernetes.io/projected/1a58e124-b1be-43ef-a54d-8baa62a26014-kube-api-access-l4s5n\") pod \"redhat-marketplace-zpcfk\" (UID: \"1a58e124-b1be-43ef-a54d-8baa62a26014\") " pod="openshift-marketplace/redhat-marketplace-zpcfk" Dec 10 19:33:21 crc kubenswrapper[4828]: I1210 19:33:21.581851 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a58e124-b1be-43ef-a54d-8baa62a26014-utilities\") pod \"redhat-marketplace-zpcfk\" (UID: \"1a58e124-b1be-43ef-a54d-8baa62a26014\") " pod="openshift-marketplace/redhat-marketplace-zpcfk" Dec 10 19:33:21 crc kubenswrapper[4828]: I1210 19:33:21.581865 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a58e124-b1be-43ef-a54d-8baa62a26014-catalog-content\") pod \"redhat-marketplace-zpcfk\" (UID: \"1a58e124-b1be-43ef-a54d-8baa62a26014\") " pod="openshift-marketplace/redhat-marketplace-zpcfk" Dec 10 19:33:21 crc kubenswrapper[4828]: I1210 19:33:21.604171 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4s5n\" (UniqueName: \"kubernetes.io/projected/1a58e124-b1be-43ef-a54d-8baa62a26014-kube-api-access-l4s5n\") pod \"redhat-marketplace-zpcfk\" (UID: \"1a58e124-b1be-43ef-a54d-8baa62a26014\") " pod="openshift-marketplace/redhat-marketplace-zpcfk" Dec 10 19:33:21 crc kubenswrapper[4828]: I1210 19:33:21.804711 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zpcfk" Dec 10 19:33:22 crc kubenswrapper[4828]: I1210 19:33:22.335095 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zpcfk"] Dec 10 19:33:22 crc kubenswrapper[4828]: I1210 19:33:22.950431 4828 generic.go:334] "Generic (PLEG): container finished" podID="1a58e124-b1be-43ef-a54d-8baa62a26014" containerID="32cb5992e7474555d0938f104d11479e1b6b9cf78f55ef6c3ea25c1825362aec" exitCode=0 Dec 10 19:33:22 crc kubenswrapper[4828]: I1210 19:33:22.950516 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zpcfk" event={"ID":"1a58e124-b1be-43ef-a54d-8baa62a26014","Type":"ContainerDied","Data":"32cb5992e7474555d0938f104d11479e1b6b9cf78f55ef6c3ea25c1825362aec"} Dec 10 19:33:22 crc kubenswrapper[4828]: I1210 19:33:22.950852 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zpcfk" event={"ID":"1a58e124-b1be-43ef-a54d-8baa62a26014","Type":"ContainerStarted","Data":"64d2dc628232bdff2c69fa86029509bc32fb797b706aebe47a65c66a28a66cf4"} Dec 10 19:33:23 crc kubenswrapper[4828]: I1210 19:33:23.963683 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zpcfk" event={"ID":"1a58e124-b1be-43ef-a54d-8baa62a26014","Type":"ContainerStarted","Data":"6f6e3127ea52483eb291833ac9bf2e773fa581bc19f0ac57141d71ceda626ed9"} Dec 10 19:33:24 crc kubenswrapper[4828]: I1210 19:33:24.975544 4828 generic.go:334] "Generic (PLEG): container finished" podID="1a58e124-b1be-43ef-a54d-8baa62a26014" containerID="6f6e3127ea52483eb291833ac9bf2e773fa581bc19f0ac57141d71ceda626ed9" exitCode=0 Dec 10 19:33:24 crc kubenswrapper[4828]: I1210 19:33:24.975583 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zpcfk" event={"ID":"1a58e124-b1be-43ef-a54d-8baa62a26014","Type":"ContainerDied","Data":"6f6e3127ea52483eb291833ac9bf2e773fa581bc19f0ac57141d71ceda626ed9"} Dec 10 19:33:26 crc kubenswrapper[4828]: I1210 19:33:26.005905 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zpcfk" event={"ID":"1a58e124-b1be-43ef-a54d-8baa62a26014","Type":"ContainerStarted","Data":"5a649a1b8e78174b02e7e991b0c31b5b92793a7b5d192ee12c710e601be9fb57"} Dec 10 19:33:26 crc kubenswrapper[4828]: I1210 19:33:26.036595 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zpcfk" podStartSLOduration=2.551059243 podStartE2EDuration="5.036556506s" podCreationTimestamp="2025-12-10 19:33:21 +0000 UTC" firstStartedPulling="2025-12-10 19:33:22.953982531 +0000 UTC m=+2283.464593546" lastFinishedPulling="2025-12-10 19:33:25.439479804 +0000 UTC m=+2285.950090809" observedRunningTime="2025-12-10 19:33:26.026506809 +0000 UTC m=+2286.537117814" watchObservedRunningTime="2025-12-10 19:33:26.036556506 +0000 UTC m=+2286.547167501" Dec 10 19:33:26 crc kubenswrapper[4828]: I1210 19:33:26.788817 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:33:26 crc kubenswrapper[4828]: E1210 19:33:26.789534 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:33:31 crc kubenswrapper[4828]: I1210 19:33:31.806328 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zpcfk" Dec 10 19:33:31 crc kubenswrapper[4828]: I1210 19:33:31.806955 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zpcfk" Dec 10 19:33:31 crc kubenswrapper[4828]: I1210 19:33:31.884240 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zpcfk" Dec 10 19:33:32 crc kubenswrapper[4828]: I1210 19:33:32.138620 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zpcfk" Dec 10 19:33:32 crc kubenswrapper[4828]: I1210 19:33:32.192682 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zpcfk"] Dec 10 19:33:34 crc kubenswrapper[4828]: I1210 19:33:34.088036 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zpcfk" podUID="1a58e124-b1be-43ef-a54d-8baa62a26014" containerName="registry-server" containerID="cri-o://5a649a1b8e78174b02e7e991b0c31b5b92793a7b5d192ee12c710e601be9fb57" gracePeriod=2 Dec 10 19:33:34 crc kubenswrapper[4828]: I1210 19:33:34.619426 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zpcfk" Dec 10 19:33:34 crc kubenswrapper[4828]: I1210 19:33:34.707861 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a58e124-b1be-43ef-a54d-8baa62a26014-utilities\") pod \"1a58e124-b1be-43ef-a54d-8baa62a26014\" (UID: \"1a58e124-b1be-43ef-a54d-8baa62a26014\") " Dec 10 19:33:34 crc kubenswrapper[4828]: I1210 19:33:34.707988 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a58e124-b1be-43ef-a54d-8baa62a26014-catalog-content\") pod \"1a58e124-b1be-43ef-a54d-8baa62a26014\" (UID: \"1a58e124-b1be-43ef-a54d-8baa62a26014\") " Dec 10 19:33:34 crc kubenswrapper[4828]: I1210 19:33:34.708023 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4s5n\" (UniqueName: \"kubernetes.io/projected/1a58e124-b1be-43ef-a54d-8baa62a26014-kube-api-access-l4s5n\") pod \"1a58e124-b1be-43ef-a54d-8baa62a26014\" (UID: \"1a58e124-b1be-43ef-a54d-8baa62a26014\") " Dec 10 19:33:34 crc kubenswrapper[4828]: I1210 19:33:34.708996 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a58e124-b1be-43ef-a54d-8baa62a26014-utilities" (OuterVolumeSpecName: "utilities") pod "1a58e124-b1be-43ef-a54d-8baa62a26014" (UID: "1a58e124-b1be-43ef-a54d-8baa62a26014"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:33:34 crc kubenswrapper[4828]: I1210 19:33:34.713561 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a58e124-b1be-43ef-a54d-8baa62a26014-kube-api-access-l4s5n" (OuterVolumeSpecName: "kube-api-access-l4s5n") pod "1a58e124-b1be-43ef-a54d-8baa62a26014" (UID: "1a58e124-b1be-43ef-a54d-8baa62a26014"). InnerVolumeSpecName "kube-api-access-l4s5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:33:34 crc kubenswrapper[4828]: I1210 19:33:34.731876 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a58e124-b1be-43ef-a54d-8baa62a26014-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a58e124-b1be-43ef-a54d-8baa62a26014" (UID: "1a58e124-b1be-43ef-a54d-8baa62a26014"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:33:34 crc kubenswrapper[4828]: I1210 19:33:34.810578 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a58e124-b1be-43ef-a54d-8baa62a26014-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:34 crc kubenswrapper[4828]: I1210 19:33:34.810614 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a58e124-b1be-43ef-a54d-8baa62a26014-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:34 crc kubenswrapper[4828]: I1210 19:33:34.810626 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4s5n\" (UniqueName: \"kubernetes.io/projected/1a58e124-b1be-43ef-a54d-8baa62a26014-kube-api-access-l4s5n\") on node \"crc\" DevicePath \"\"" Dec 10 19:33:35 crc kubenswrapper[4828]: I1210 19:33:35.099912 4828 generic.go:334] "Generic (PLEG): container finished" podID="1a58e124-b1be-43ef-a54d-8baa62a26014" containerID="5a649a1b8e78174b02e7e991b0c31b5b92793a7b5d192ee12c710e601be9fb57" exitCode=0 Dec 10 19:33:35 crc kubenswrapper[4828]: I1210 19:33:35.100005 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zpcfk" Dec 10 19:33:35 crc kubenswrapper[4828]: I1210 19:33:35.100004 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zpcfk" event={"ID":"1a58e124-b1be-43ef-a54d-8baa62a26014","Type":"ContainerDied","Data":"5a649a1b8e78174b02e7e991b0c31b5b92793a7b5d192ee12c710e601be9fb57"} Dec 10 19:33:35 crc kubenswrapper[4828]: I1210 19:33:35.100360 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zpcfk" event={"ID":"1a58e124-b1be-43ef-a54d-8baa62a26014","Type":"ContainerDied","Data":"64d2dc628232bdff2c69fa86029509bc32fb797b706aebe47a65c66a28a66cf4"} Dec 10 19:33:35 crc kubenswrapper[4828]: I1210 19:33:35.100380 4828 scope.go:117] "RemoveContainer" containerID="5a649a1b8e78174b02e7e991b0c31b5b92793a7b5d192ee12c710e601be9fb57" Dec 10 19:33:35 crc kubenswrapper[4828]: I1210 19:33:35.122546 4828 scope.go:117] "RemoveContainer" containerID="6f6e3127ea52483eb291833ac9bf2e773fa581bc19f0ac57141d71ceda626ed9" Dec 10 19:33:35 crc kubenswrapper[4828]: I1210 19:33:35.139969 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zpcfk"] Dec 10 19:33:35 crc kubenswrapper[4828]: I1210 19:33:35.150311 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zpcfk"] Dec 10 19:33:35 crc kubenswrapper[4828]: I1210 19:33:35.196736 4828 scope.go:117] "RemoveContainer" containerID="32cb5992e7474555d0938f104d11479e1b6b9cf78f55ef6c3ea25c1825362aec" Dec 10 19:33:35 crc kubenswrapper[4828]: I1210 19:33:35.228954 4828 scope.go:117] "RemoveContainer" containerID="5a649a1b8e78174b02e7e991b0c31b5b92793a7b5d192ee12c710e601be9fb57" Dec 10 19:33:35 crc kubenswrapper[4828]: E1210 19:33:35.229419 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a649a1b8e78174b02e7e991b0c31b5b92793a7b5d192ee12c710e601be9fb57\": container with ID starting with 5a649a1b8e78174b02e7e991b0c31b5b92793a7b5d192ee12c710e601be9fb57 not found: ID does not exist" containerID="5a649a1b8e78174b02e7e991b0c31b5b92793a7b5d192ee12c710e601be9fb57" Dec 10 19:33:35 crc kubenswrapper[4828]: I1210 19:33:35.229444 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a649a1b8e78174b02e7e991b0c31b5b92793a7b5d192ee12c710e601be9fb57"} err="failed to get container status \"5a649a1b8e78174b02e7e991b0c31b5b92793a7b5d192ee12c710e601be9fb57\": rpc error: code = NotFound desc = could not find container \"5a649a1b8e78174b02e7e991b0c31b5b92793a7b5d192ee12c710e601be9fb57\": container with ID starting with 5a649a1b8e78174b02e7e991b0c31b5b92793a7b5d192ee12c710e601be9fb57 not found: ID does not exist" Dec 10 19:33:35 crc kubenswrapper[4828]: I1210 19:33:35.229465 4828 scope.go:117] "RemoveContainer" containerID="6f6e3127ea52483eb291833ac9bf2e773fa581bc19f0ac57141d71ceda626ed9" Dec 10 19:33:35 crc kubenswrapper[4828]: E1210 19:33:35.230643 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f6e3127ea52483eb291833ac9bf2e773fa581bc19f0ac57141d71ceda626ed9\": container with ID starting with 6f6e3127ea52483eb291833ac9bf2e773fa581bc19f0ac57141d71ceda626ed9 not found: ID does not exist" containerID="6f6e3127ea52483eb291833ac9bf2e773fa581bc19f0ac57141d71ceda626ed9" Dec 10 19:33:35 crc kubenswrapper[4828]: I1210 19:33:35.230665 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f6e3127ea52483eb291833ac9bf2e773fa581bc19f0ac57141d71ceda626ed9"} err="failed to get container status \"6f6e3127ea52483eb291833ac9bf2e773fa581bc19f0ac57141d71ceda626ed9\": rpc error: code = NotFound desc = could not find container \"6f6e3127ea52483eb291833ac9bf2e773fa581bc19f0ac57141d71ceda626ed9\": container with ID starting with 6f6e3127ea52483eb291833ac9bf2e773fa581bc19f0ac57141d71ceda626ed9 not found: ID does not exist" Dec 10 19:33:35 crc kubenswrapper[4828]: I1210 19:33:35.230680 4828 scope.go:117] "RemoveContainer" containerID="32cb5992e7474555d0938f104d11479e1b6b9cf78f55ef6c3ea25c1825362aec" Dec 10 19:33:35 crc kubenswrapper[4828]: E1210 19:33:35.230925 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32cb5992e7474555d0938f104d11479e1b6b9cf78f55ef6c3ea25c1825362aec\": container with ID starting with 32cb5992e7474555d0938f104d11479e1b6b9cf78f55ef6c3ea25c1825362aec not found: ID does not exist" containerID="32cb5992e7474555d0938f104d11479e1b6b9cf78f55ef6c3ea25c1825362aec" Dec 10 19:33:35 crc kubenswrapper[4828]: I1210 19:33:35.230969 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32cb5992e7474555d0938f104d11479e1b6b9cf78f55ef6c3ea25c1825362aec"} err="failed to get container status \"32cb5992e7474555d0938f104d11479e1b6b9cf78f55ef6c3ea25c1825362aec\": rpc error: code = NotFound desc = could not find container \"32cb5992e7474555d0938f104d11479e1b6b9cf78f55ef6c3ea25c1825362aec\": container with ID starting with 32cb5992e7474555d0938f104d11479e1b6b9cf78f55ef6c3ea25c1825362aec not found: ID does not exist" Dec 10 19:33:35 crc kubenswrapper[4828]: I1210 19:33:35.812339 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a58e124-b1be-43ef-a54d-8baa62a26014" path="/var/lib/kubelet/pods/1a58e124-b1be-43ef-a54d-8baa62a26014/volumes" Dec 10 19:33:39 crc kubenswrapper[4828]: I1210 19:33:39.797253 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:33:39 crc kubenswrapper[4828]: E1210 19:33:39.798299 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:33:41 crc kubenswrapper[4828]: I1210 19:33:41.633713 4828 scope.go:117] "RemoveContainer" containerID="069149e5295320fff42b4e6b3c3d6f2c88abc48d4b319290aca58392c560dc3b" Dec 10 19:33:51 crc kubenswrapper[4828]: I1210 19:33:51.789712 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:33:51 crc kubenswrapper[4828]: E1210 19:33:51.790604 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:33:58 crc kubenswrapper[4828]: I1210 19:33:58.381968 4828 generic.go:334] "Generic (PLEG): container finished" podID="b3f83864-18ee-4493-b420-a536371c509b" containerID="244e4e284e09618028876df9e39983135fe5f1fdea0b0f4bf99441dbd21808b7" exitCode=0 Dec 10 19:33:58 crc kubenswrapper[4828]: I1210 19:33:58.382206 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" event={"ID":"b3f83864-18ee-4493-b420-a536371c509b","Type":"ContainerDied","Data":"244e4e284e09618028876df9e39983135fe5f1fdea0b0f4bf99441dbd21808b7"} Dec 10 19:33:59 crc kubenswrapper[4828]: I1210 19:33:59.836975 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" Dec 10 19:33:59 crc kubenswrapper[4828]: I1210 19:33:59.998626 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3f83864-18ee-4493-b420-a536371c509b-ssh-key\") pod \"b3f83864-18ee-4493-b420-a536371c509b\" (UID: \"b3f83864-18ee-4493-b420-a536371c509b\") " Dec 10 19:33:59 crc kubenswrapper[4828]: I1210 19:33:59.999170 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3f83864-18ee-4493-b420-a536371c509b-inventory\") pod \"b3f83864-18ee-4493-b420-a536371c509b\" (UID: \"b3f83864-18ee-4493-b420-a536371c509b\") " Dec 10 19:33:59 crc kubenswrapper[4828]: I1210 19:33:59.999207 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2vkd\" (UniqueName: \"kubernetes.io/projected/b3f83864-18ee-4493-b420-a536371c509b-kube-api-access-j2vkd\") pod \"b3f83864-18ee-4493-b420-a536371c509b\" (UID: \"b3f83864-18ee-4493-b420-a536371c509b\") " Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.003813 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3f83864-18ee-4493-b420-a536371c509b-kube-api-access-j2vkd" (OuterVolumeSpecName: "kube-api-access-j2vkd") pod "b3f83864-18ee-4493-b420-a536371c509b" (UID: "b3f83864-18ee-4493-b420-a536371c509b"). InnerVolumeSpecName "kube-api-access-j2vkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.030112 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3f83864-18ee-4493-b420-a536371c509b-inventory" (OuterVolumeSpecName: "inventory") pod "b3f83864-18ee-4493-b420-a536371c509b" (UID: "b3f83864-18ee-4493-b420-a536371c509b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.036408 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3f83864-18ee-4493-b420-a536371c509b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b3f83864-18ee-4493-b420-a536371c509b" (UID: "b3f83864-18ee-4493-b420-a536371c509b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.102844 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2vkd\" (UniqueName: \"kubernetes.io/projected/b3f83864-18ee-4493-b420-a536371c509b-kube-api-access-j2vkd\") on node \"crc\" DevicePath \"\"" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.102908 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3f83864-18ee-4493-b420-a536371c509b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.102931 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3f83864-18ee-4493-b420-a536371c509b-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.404596 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" event={"ID":"b3f83864-18ee-4493-b420-a536371c509b","Type":"ContainerDied","Data":"35310d7018e2385a55a9ff474da62bc1b326dac247672d30e8058801a1c76547"} Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.404658 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35310d7018e2385a55a9ff474da62bc1b326dac247672d30e8058801a1c76547" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.404741 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xd599" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.494722 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cm8mx"] Dec 10 19:34:00 crc kubenswrapper[4828]: E1210 19:34:00.495412 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3f83864-18ee-4493-b420-a536371c509b" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.495438 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3f83864-18ee-4493-b420-a536371c509b" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:34:00 crc kubenswrapper[4828]: E1210 19:34:00.495460 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a58e124-b1be-43ef-a54d-8baa62a26014" containerName="extract-utilities" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.495471 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a58e124-b1be-43ef-a54d-8baa62a26014" containerName="extract-utilities" Dec 10 19:34:00 crc kubenswrapper[4828]: E1210 19:34:00.495512 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a58e124-b1be-43ef-a54d-8baa62a26014" containerName="extract-content" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.495519 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a58e124-b1be-43ef-a54d-8baa62a26014" containerName="extract-content" Dec 10 19:34:00 crc kubenswrapper[4828]: E1210 19:34:00.495529 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a58e124-b1be-43ef-a54d-8baa62a26014" containerName="registry-server" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.495536 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a58e124-b1be-43ef-a54d-8baa62a26014" containerName="registry-server" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.495834 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a58e124-b1be-43ef-a54d-8baa62a26014" containerName="registry-server" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.495852 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3f83864-18ee-4493-b420-a536371c509b" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.496765 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.500080 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.500206 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.509229 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.509267 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.520147 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cm8mx"] Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.613297 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxbw7\" (UniqueName: \"kubernetes.io/projected/cba6fd50-fd63-4672-8d84-396aba02ff04-kube-api-access-zxbw7\") pod \"ssh-known-hosts-edpm-deployment-cm8mx\" (UID: \"cba6fd50-fd63-4672-8d84-396aba02ff04\") " pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.613565 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cba6fd50-fd63-4672-8d84-396aba02ff04-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cm8mx\" (UID: \"cba6fd50-fd63-4672-8d84-396aba02ff04\") " pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.613845 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cba6fd50-fd63-4672-8d84-396aba02ff04-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cm8mx\" (UID: \"cba6fd50-fd63-4672-8d84-396aba02ff04\") " pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.715556 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cba6fd50-fd63-4672-8d84-396aba02ff04-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cm8mx\" (UID: \"cba6fd50-fd63-4672-8d84-396aba02ff04\") " pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.715685 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cba6fd50-fd63-4672-8d84-396aba02ff04-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cm8mx\" (UID: \"cba6fd50-fd63-4672-8d84-396aba02ff04\") " pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.715818 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxbw7\" (UniqueName: \"kubernetes.io/projected/cba6fd50-fd63-4672-8d84-396aba02ff04-kube-api-access-zxbw7\") pod \"ssh-known-hosts-edpm-deployment-cm8mx\" (UID: \"cba6fd50-fd63-4672-8d84-396aba02ff04\") " pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.719684 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cba6fd50-fd63-4672-8d84-396aba02ff04-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-cm8mx\" (UID: \"cba6fd50-fd63-4672-8d84-396aba02ff04\") " pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.720142 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cba6fd50-fd63-4672-8d84-396aba02ff04-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-cm8mx\" (UID: \"cba6fd50-fd63-4672-8d84-396aba02ff04\") " pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.732903 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxbw7\" (UniqueName: \"kubernetes.io/projected/cba6fd50-fd63-4672-8d84-396aba02ff04-kube-api-access-zxbw7\") pod \"ssh-known-hosts-edpm-deployment-cm8mx\" (UID: \"cba6fd50-fd63-4672-8d84-396aba02ff04\") " pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" Dec 10 19:34:00 crc kubenswrapper[4828]: I1210 19:34:00.818605 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" Dec 10 19:34:01 crc kubenswrapper[4828]: I1210 19:34:01.366924 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-cm8mx"] Dec 10 19:34:01 crc kubenswrapper[4828]: I1210 19:34:01.370610 4828 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:34:01 crc kubenswrapper[4828]: I1210 19:34:01.415616 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" event={"ID":"cba6fd50-fd63-4672-8d84-396aba02ff04","Type":"ContainerStarted","Data":"d6fbc2d8c712511852778292758ac9ffcf75e67ea8a8d93ada7e7ccde80a3955"} Dec 10 19:34:02 crc kubenswrapper[4828]: I1210 19:34:02.437178 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" event={"ID":"cba6fd50-fd63-4672-8d84-396aba02ff04","Type":"ContainerStarted","Data":"ea2bf47f39d9a72111b5d74deedd7123bcce6b95e90fb116b383201a03e0d16d"} Dec 10 19:34:02 crc kubenswrapper[4828]: I1210 19:34:02.458136 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" podStartSLOduration=1.8597312270000002 podStartE2EDuration="2.458114262s" podCreationTimestamp="2025-12-10 19:34:00 +0000 UTC" firstStartedPulling="2025-12-10 19:34:01.370377533 +0000 UTC m=+2321.880988548" lastFinishedPulling="2025-12-10 19:34:01.968760538 +0000 UTC m=+2322.479371583" observedRunningTime="2025-12-10 19:34:02.453041977 +0000 UTC m=+2322.963653002" watchObservedRunningTime="2025-12-10 19:34:02.458114262 +0000 UTC m=+2322.968725277" Dec 10 19:34:02 crc kubenswrapper[4828]: I1210 19:34:02.788570 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:34:02 crc kubenswrapper[4828]: E1210 19:34:02.788927 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:34:09 crc kubenswrapper[4828]: I1210 19:34:09.515742 4828 generic.go:334] "Generic (PLEG): container finished" podID="cba6fd50-fd63-4672-8d84-396aba02ff04" containerID="ea2bf47f39d9a72111b5d74deedd7123bcce6b95e90fb116b383201a03e0d16d" exitCode=0 Dec 10 19:34:09 crc kubenswrapper[4828]: I1210 19:34:09.515851 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" event={"ID":"cba6fd50-fd63-4672-8d84-396aba02ff04","Type":"ContainerDied","Data":"ea2bf47f39d9a72111b5d74deedd7123bcce6b95e90fb116b383201a03e0d16d"} Dec 10 19:34:10 crc kubenswrapper[4828]: I1210 19:34:10.965152 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.067828 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cba6fd50-fd63-4672-8d84-396aba02ff04-ssh-key-openstack-edpm-ipam\") pod \"cba6fd50-fd63-4672-8d84-396aba02ff04\" (UID: \"cba6fd50-fd63-4672-8d84-396aba02ff04\") " Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.067934 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxbw7\" (UniqueName: \"kubernetes.io/projected/cba6fd50-fd63-4672-8d84-396aba02ff04-kube-api-access-zxbw7\") pod \"cba6fd50-fd63-4672-8d84-396aba02ff04\" (UID: \"cba6fd50-fd63-4672-8d84-396aba02ff04\") " Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.068128 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cba6fd50-fd63-4672-8d84-396aba02ff04-inventory-0\") pod \"cba6fd50-fd63-4672-8d84-396aba02ff04\" (UID: \"cba6fd50-fd63-4672-8d84-396aba02ff04\") " Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.074278 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cba6fd50-fd63-4672-8d84-396aba02ff04-kube-api-access-zxbw7" (OuterVolumeSpecName: "kube-api-access-zxbw7") pod "cba6fd50-fd63-4672-8d84-396aba02ff04" (UID: "cba6fd50-fd63-4672-8d84-396aba02ff04"). InnerVolumeSpecName "kube-api-access-zxbw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.105997 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cba6fd50-fd63-4672-8d84-396aba02ff04-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "cba6fd50-fd63-4672-8d84-396aba02ff04" (UID: "cba6fd50-fd63-4672-8d84-396aba02ff04"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.111596 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cba6fd50-fd63-4672-8d84-396aba02ff04-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "cba6fd50-fd63-4672-8d84-396aba02ff04" (UID: "cba6fd50-fd63-4672-8d84-396aba02ff04"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.172379 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cba6fd50-fd63-4672-8d84-396aba02ff04-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.172437 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxbw7\" (UniqueName: \"kubernetes.io/projected/cba6fd50-fd63-4672-8d84-396aba02ff04-kube-api-access-zxbw7\") on node \"crc\" DevicePath \"\"" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.172453 4828 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cba6fd50-fd63-4672-8d84-396aba02ff04-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.538759 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" event={"ID":"cba6fd50-fd63-4672-8d84-396aba02ff04","Type":"ContainerDied","Data":"d6fbc2d8c712511852778292758ac9ffcf75e67ea8a8d93ada7e7ccde80a3955"} Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.539255 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6fbc2d8c712511852778292758ac9ffcf75e67ea8a8d93ada7e7ccde80a3955" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.538829 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-cm8mx" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.606366 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg"] Dec 10 19:34:11 crc kubenswrapper[4828]: E1210 19:34:11.606796 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cba6fd50-fd63-4672-8d84-396aba02ff04" containerName="ssh-known-hosts-edpm-deployment" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.606829 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="cba6fd50-fd63-4672-8d84-396aba02ff04" containerName="ssh-known-hosts-edpm-deployment" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.607073 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="cba6fd50-fd63-4672-8d84-396aba02ff04" containerName="ssh-known-hosts-edpm-deployment" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.607952 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.609901 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.610041 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.610884 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.612054 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.616810 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg"] Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.788738 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3bac884-e92c-44ec-999f-1e8ab80f4bca-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wzcbg\" (UID: \"d3bac884-e92c-44ec-999f-1e8ab80f4bca\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.789128 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3bac884-e92c-44ec-999f-1e8ab80f4bca-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wzcbg\" (UID: \"d3bac884-e92c-44ec-999f-1e8ab80f4bca\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.789168 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvrhl\" (UniqueName: \"kubernetes.io/projected/d3bac884-e92c-44ec-999f-1e8ab80f4bca-kube-api-access-dvrhl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wzcbg\" (UID: \"d3bac884-e92c-44ec-999f-1e8ab80f4bca\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.890635 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3bac884-e92c-44ec-999f-1e8ab80f4bca-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wzcbg\" (UID: \"d3bac884-e92c-44ec-999f-1e8ab80f4bca\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.890790 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3bac884-e92c-44ec-999f-1e8ab80f4bca-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wzcbg\" (UID: \"d3bac884-e92c-44ec-999f-1e8ab80f4bca\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.890834 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvrhl\" (UniqueName: \"kubernetes.io/projected/d3bac884-e92c-44ec-999f-1e8ab80f4bca-kube-api-access-dvrhl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wzcbg\" (UID: \"d3bac884-e92c-44ec-999f-1e8ab80f4bca\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.897715 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3bac884-e92c-44ec-999f-1e8ab80f4bca-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wzcbg\" (UID: \"d3bac884-e92c-44ec-999f-1e8ab80f4bca\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.897809 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3bac884-e92c-44ec-999f-1e8ab80f4bca-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wzcbg\" (UID: \"d3bac884-e92c-44ec-999f-1e8ab80f4bca\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.912998 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvrhl\" (UniqueName: \"kubernetes.io/projected/d3bac884-e92c-44ec-999f-1e8ab80f4bca-kube-api-access-dvrhl\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wzcbg\" (UID: \"d3bac884-e92c-44ec-999f-1e8ab80f4bca\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" Dec 10 19:34:11 crc kubenswrapper[4828]: I1210 19:34:11.928282 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" Dec 10 19:34:12 crc kubenswrapper[4828]: I1210 19:34:12.456586 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg"] Dec 10 19:34:12 crc kubenswrapper[4828]: I1210 19:34:12.552042 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" event={"ID":"d3bac884-e92c-44ec-999f-1e8ab80f4bca","Type":"ContainerStarted","Data":"ff4918ef2ca9908025e22c76a59f46138ded280388c7ffd0b641ba1894a6dc2a"} Dec 10 19:34:13 crc kubenswrapper[4828]: I1210 19:34:13.563198 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" event={"ID":"d3bac884-e92c-44ec-999f-1e8ab80f4bca","Type":"ContainerStarted","Data":"8bc9527e3a65166ea0e7b92b2c75e36d8984f0b9f474aa22d014035080be4927"} Dec 10 19:34:13 crc kubenswrapper[4828]: I1210 19:34:13.583975 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" podStartSLOduration=1.771669022 podStartE2EDuration="2.583956924s" podCreationTimestamp="2025-12-10 19:34:11 +0000 UTC" firstStartedPulling="2025-12-10 19:34:12.458752286 +0000 UTC m=+2332.969363291" lastFinishedPulling="2025-12-10 19:34:13.271040188 +0000 UTC m=+2333.781651193" observedRunningTime="2025-12-10 19:34:13.577130612 +0000 UTC m=+2334.087741627" watchObservedRunningTime="2025-12-10 19:34:13.583956924 +0000 UTC m=+2334.094567929" Dec 10 19:34:13 crc kubenswrapper[4828]: I1210 19:34:13.788672 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:34:13 crc kubenswrapper[4828]: E1210 19:34:13.790698 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:34:21 crc kubenswrapper[4828]: I1210 19:34:21.647657 4828 generic.go:334] "Generic (PLEG): container finished" podID="d3bac884-e92c-44ec-999f-1e8ab80f4bca" containerID="8bc9527e3a65166ea0e7b92b2c75e36d8984f0b9f474aa22d014035080be4927" exitCode=0 Dec 10 19:34:21 crc kubenswrapper[4828]: I1210 19:34:21.647740 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" event={"ID":"d3bac884-e92c-44ec-999f-1e8ab80f4bca","Type":"ContainerDied","Data":"8bc9527e3a65166ea0e7b92b2c75e36d8984f0b9f474aa22d014035080be4927"} Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.127781 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.265636 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3bac884-e92c-44ec-999f-1e8ab80f4bca-inventory\") pod \"d3bac884-e92c-44ec-999f-1e8ab80f4bca\" (UID: \"d3bac884-e92c-44ec-999f-1e8ab80f4bca\") " Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.265933 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvrhl\" (UniqueName: \"kubernetes.io/projected/d3bac884-e92c-44ec-999f-1e8ab80f4bca-kube-api-access-dvrhl\") pod \"d3bac884-e92c-44ec-999f-1e8ab80f4bca\" (UID: \"d3bac884-e92c-44ec-999f-1e8ab80f4bca\") " Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.266058 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3bac884-e92c-44ec-999f-1e8ab80f4bca-ssh-key\") pod \"d3bac884-e92c-44ec-999f-1e8ab80f4bca\" (UID: \"d3bac884-e92c-44ec-999f-1e8ab80f4bca\") " Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.271474 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3bac884-e92c-44ec-999f-1e8ab80f4bca-kube-api-access-dvrhl" (OuterVolumeSpecName: "kube-api-access-dvrhl") pod "d3bac884-e92c-44ec-999f-1e8ab80f4bca" (UID: "d3bac884-e92c-44ec-999f-1e8ab80f4bca"). InnerVolumeSpecName "kube-api-access-dvrhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.302366 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3bac884-e92c-44ec-999f-1e8ab80f4bca-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d3bac884-e92c-44ec-999f-1e8ab80f4bca" (UID: "d3bac884-e92c-44ec-999f-1e8ab80f4bca"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.304019 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3bac884-e92c-44ec-999f-1e8ab80f4bca-inventory" (OuterVolumeSpecName: "inventory") pod "d3bac884-e92c-44ec-999f-1e8ab80f4bca" (UID: "d3bac884-e92c-44ec-999f-1e8ab80f4bca"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.368763 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvrhl\" (UniqueName: \"kubernetes.io/projected/d3bac884-e92c-44ec-999f-1e8ab80f4bca-kube-api-access-dvrhl\") on node \"crc\" DevicePath \"\"" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.368821 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d3bac884-e92c-44ec-999f-1e8ab80f4bca-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.368837 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d3bac884-e92c-44ec-999f-1e8ab80f4bca-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.668190 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" event={"ID":"d3bac884-e92c-44ec-999f-1e8ab80f4bca","Type":"ContainerDied","Data":"ff4918ef2ca9908025e22c76a59f46138ded280388c7ffd0b641ba1894a6dc2a"} Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.668237 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff4918ef2ca9908025e22c76a59f46138ded280388c7ffd0b641ba1894a6dc2a" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.668290 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wzcbg" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.744726 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9"] Dec 10 19:34:23 crc kubenswrapper[4828]: E1210 19:34:23.745399 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3bac884-e92c-44ec-999f-1e8ab80f4bca" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.745436 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3bac884-e92c-44ec-999f-1e8ab80f4bca" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.745712 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3bac884-e92c-44ec-999f-1e8ab80f4bca" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.746829 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.748742 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.749392 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.750302 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.753626 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.755582 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9"] Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.879174 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9\" (UID: \"c74d14a7-f5e1-40cc-bde7-72d19e285ae9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.879460 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwwq7\" (UniqueName: \"kubernetes.io/projected/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-kube-api-access-cwwq7\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9\" (UID: \"c74d14a7-f5e1-40cc-bde7-72d19e285ae9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.879553 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9\" (UID: \"c74d14a7-f5e1-40cc-bde7-72d19e285ae9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.981820 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9\" (UID: \"c74d14a7-f5e1-40cc-bde7-72d19e285ae9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.981946 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwwq7\" (UniqueName: \"kubernetes.io/projected/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-kube-api-access-cwwq7\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9\" (UID: \"c74d14a7-f5e1-40cc-bde7-72d19e285ae9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.981983 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9\" (UID: \"c74d14a7-f5e1-40cc-bde7-72d19e285ae9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.986512 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9\" (UID: \"c74d14a7-f5e1-40cc-bde7-72d19e285ae9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" Dec 10 19:34:23 crc kubenswrapper[4828]: I1210 19:34:23.986557 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9\" (UID: \"c74d14a7-f5e1-40cc-bde7-72d19e285ae9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" Dec 10 19:34:24 crc kubenswrapper[4828]: I1210 19:34:24.018235 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwwq7\" (UniqueName: \"kubernetes.io/projected/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-kube-api-access-cwwq7\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9\" (UID: \"c74d14a7-f5e1-40cc-bde7-72d19e285ae9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" Dec 10 19:34:24 crc kubenswrapper[4828]: I1210 19:34:24.067391 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" Dec 10 19:34:24 crc kubenswrapper[4828]: I1210 19:34:24.614606 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9"] Dec 10 19:34:24 crc kubenswrapper[4828]: I1210 19:34:24.678785 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" event={"ID":"c74d14a7-f5e1-40cc-bde7-72d19e285ae9","Type":"ContainerStarted","Data":"60aa7f320c9a2642596522994f027db0355299ecc4c50c7725a605c6032e08a0"} Dec 10 19:34:25 crc kubenswrapper[4828]: I1210 19:34:25.696790 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" event={"ID":"c74d14a7-f5e1-40cc-bde7-72d19e285ae9","Type":"ContainerStarted","Data":"1f632ced838f5a5d8a70820190d43eccf97480293937f828230464349cf64aab"} Dec 10 19:34:25 crc kubenswrapper[4828]: I1210 19:34:25.718886 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" podStartSLOduration=2.095477763 podStartE2EDuration="2.718859308s" podCreationTimestamp="2025-12-10 19:34:23 +0000 UTC" firstStartedPulling="2025-12-10 19:34:24.625005365 +0000 UTC m=+2345.135616370" lastFinishedPulling="2025-12-10 19:34:25.24838691 +0000 UTC m=+2345.758997915" observedRunningTime="2025-12-10 19:34:25.711178754 +0000 UTC m=+2346.221789769" watchObservedRunningTime="2025-12-10 19:34:25.718859308 +0000 UTC m=+2346.229470323" Dec 10 19:34:28 crc kubenswrapper[4828]: I1210 19:34:28.789603 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:34:28 crc kubenswrapper[4828]: E1210 19:34:28.790458 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:34:34 crc kubenswrapper[4828]: I1210 19:34:34.796268 4828 generic.go:334] "Generic (PLEG): container finished" podID="c74d14a7-f5e1-40cc-bde7-72d19e285ae9" containerID="1f632ced838f5a5d8a70820190d43eccf97480293937f828230464349cf64aab" exitCode=0 Dec 10 19:34:34 crc kubenswrapper[4828]: I1210 19:34:34.796354 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" event={"ID":"c74d14a7-f5e1-40cc-bde7-72d19e285ae9","Type":"ContainerDied","Data":"1f632ced838f5a5d8a70820190d43eccf97480293937f828230464349cf64aab"} Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.273076 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.375126 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwwq7\" (UniqueName: \"kubernetes.io/projected/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-kube-api-access-cwwq7\") pod \"c74d14a7-f5e1-40cc-bde7-72d19e285ae9\" (UID: \"c74d14a7-f5e1-40cc-bde7-72d19e285ae9\") " Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.375167 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-inventory\") pod \"c74d14a7-f5e1-40cc-bde7-72d19e285ae9\" (UID: \"c74d14a7-f5e1-40cc-bde7-72d19e285ae9\") " Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.375381 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-ssh-key\") pod \"c74d14a7-f5e1-40cc-bde7-72d19e285ae9\" (UID: \"c74d14a7-f5e1-40cc-bde7-72d19e285ae9\") " Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.382280 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-kube-api-access-cwwq7" (OuterVolumeSpecName: "kube-api-access-cwwq7") pod "c74d14a7-f5e1-40cc-bde7-72d19e285ae9" (UID: "c74d14a7-f5e1-40cc-bde7-72d19e285ae9"). InnerVolumeSpecName "kube-api-access-cwwq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.415111 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c74d14a7-f5e1-40cc-bde7-72d19e285ae9" (UID: "c74d14a7-f5e1-40cc-bde7-72d19e285ae9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.429666 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-inventory" (OuterVolumeSpecName: "inventory") pod "c74d14a7-f5e1-40cc-bde7-72d19e285ae9" (UID: "c74d14a7-f5e1-40cc-bde7-72d19e285ae9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.477582 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.477633 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwwq7\" (UniqueName: \"kubernetes.io/projected/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-kube-api-access-cwwq7\") on node \"crc\" DevicePath \"\"" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.477649 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c74d14a7-f5e1-40cc-bde7-72d19e285ae9-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.817877 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" event={"ID":"c74d14a7-f5e1-40cc-bde7-72d19e285ae9","Type":"ContainerDied","Data":"60aa7f320c9a2642596522994f027db0355299ecc4c50c7725a605c6032e08a0"} Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.817925 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60aa7f320c9a2642596522994f027db0355299ecc4c50c7725a605c6032e08a0" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.817962 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.914840 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv"] Dec 10 19:34:36 crc kubenswrapper[4828]: E1210 19:34:36.915541 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c74d14a7-f5e1-40cc-bde7-72d19e285ae9" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.915619 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c74d14a7-f5e1-40cc-bde7-72d19e285ae9" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.916002 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c74d14a7-f5e1-40cc-bde7-72d19e285ae9" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.916969 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.919560 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.919726 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.920020 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.920091 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.920253 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.920311 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.920768 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.921459 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.921733 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.925790 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv"] Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.991918 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.991965 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.992012 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.992052 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.992190 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.992250 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.992298 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.992483 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.992558 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.992669 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.992718 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.992768 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c878l\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-kube-api-access-c878l\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.992875 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.992970 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.993019 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:36 crc kubenswrapper[4828]: I1210 19:34:36.993226 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.095450 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.095535 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.095556 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.095609 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.095639 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.095658 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.095677 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.095698 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.095738 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.095765 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.095809 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.095829 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.095851 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c878l\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-kube-api-access-c878l\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.095871 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.095901 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.095930 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.100630 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.100639 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.100639 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.100669 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.101438 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.101530 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.102894 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.103611 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.103906 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.104961 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.105101 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.105168 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.105688 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.107285 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.108432 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.119982 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c878l\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-kube-api-access-c878l\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.247245 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:34:37 crc kubenswrapper[4828]: I1210 19:34:37.932837 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv"] Dec 10 19:34:38 crc kubenswrapper[4828]: I1210 19:34:38.837300 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" event={"ID":"5726d3d2-c797-4fca-aa6c-8aec98c6061b","Type":"ContainerStarted","Data":"9d6bfed25e610e855bc94ca90deb47d28aed09a98803a8fba541cf5ac974afbd"} Dec 10 19:34:39 crc kubenswrapper[4828]: I1210 19:34:39.846949 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" event={"ID":"5726d3d2-c797-4fca-aa6c-8aec98c6061b","Type":"ContainerStarted","Data":"049485cfdbe4e24fe1d11b8c1ee03f62ba5e91996a24354ecc095cf0dc29c76c"} Dec 10 19:34:39 crc kubenswrapper[4828]: I1210 19:34:39.874007 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" podStartSLOduration=3.156865722 podStartE2EDuration="3.873984651s" podCreationTimestamp="2025-12-10 19:34:36 +0000 UTC" firstStartedPulling="2025-12-10 19:34:37.935333681 +0000 UTC m=+2358.445944686" lastFinishedPulling="2025-12-10 19:34:38.6524526 +0000 UTC m=+2359.163063615" observedRunningTime="2025-12-10 19:34:39.868272379 +0000 UTC m=+2360.378883384" watchObservedRunningTime="2025-12-10 19:34:39.873984651 +0000 UTC m=+2360.384595656" Dec 10 19:34:42 crc kubenswrapper[4828]: I1210 19:34:42.789957 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:34:42 crc kubenswrapper[4828]: E1210 19:34:42.791084 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:34:55 crc kubenswrapper[4828]: I1210 19:34:55.073744 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-2n2fz"] Dec 10 19:34:55 crc kubenswrapper[4828]: I1210 19:34:55.089253 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-2n2fz"] Dec 10 19:34:55 crc kubenswrapper[4828]: I1210 19:34:55.789657 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:34:55 crc kubenswrapper[4828]: E1210 19:34:55.790012 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:34:55 crc kubenswrapper[4828]: I1210 19:34:55.801501 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2449558a-d727-461c-acb2-99e380868fb1" path="/var/lib/kubelet/pods/2449558a-d727-461c-acb2-99e380868fb1/volumes" Dec 10 19:35:08 crc kubenswrapper[4828]: I1210 19:35:08.766718 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l2pvb"] Dec 10 19:35:08 crc kubenswrapper[4828]: I1210 19:35:08.770356 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l2pvb" Dec 10 19:35:08 crc kubenswrapper[4828]: I1210 19:35:08.793978 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l2pvb"] Dec 10 19:35:08 crc kubenswrapper[4828]: I1210 19:35:08.796185 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:35:08 crc kubenswrapper[4828]: E1210 19:35:08.796505 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:35:08 crc kubenswrapper[4828]: I1210 19:35:08.929188 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e89f69df-9d96-4d7e-be8a-b32d3989de95-catalog-content\") pod \"community-operators-l2pvb\" (UID: \"e89f69df-9d96-4d7e-be8a-b32d3989de95\") " pod="openshift-marketplace/community-operators-l2pvb" Dec 10 19:35:08 crc kubenswrapper[4828]: I1210 19:35:08.929380 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jswb4\" (UniqueName: \"kubernetes.io/projected/e89f69df-9d96-4d7e-be8a-b32d3989de95-kube-api-access-jswb4\") pod \"community-operators-l2pvb\" (UID: \"e89f69df-9d96-4d7e-be8a-b32d3989de95\") " pod="openshift-marketplace/community-operators-l2pvb" Dec 10 19:35:08 crc kubenswrapper[4828]: I1210 19:35:08.929483 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e89f69df-9d96-4d7e-be8a-b32d3989de95-utilities\") pod \"community-operators-l2pvb\" (UID: \"e89f69df-9d96-4d7e-be8a-b32d3989de95\") " pod="openshift-marketplace/community-operators-l2pvb" Dec 10 19:35:09 crc kubenswrapper[4828]: I1210 19:35:09.032738 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e89f69df-9d96-4d7e-be8a-b32d3989de95-catalog-content\") pod \"community-operators-l2pvb\" (UID: \"e89f69df-9d96-4d7e-be8a-b32d3989de95\") " pod="openshift-marketplace/community-operators-l2pvb" Dec 10 19:35:09 crc kubenswrapper[4828]: I1210 19:35:09.032901 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jswb4\" (UniqueName: \"kubernetes.io/projected/e89f69df-9d96-4d7e-be8a-b32d3989de95-kube-api-access-jswb4\") pod \"community-operators-l2pvb\" (UID: \"e89f69df-9d96-4d7e-be8a-b32d3989de95\") " pod="openshift-marketplace/community-operators-l2pvb" Dec 10 19:35:09 crc kubenswrapper[4828]: I1210 19:35:09.032957 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e89f69df-9d96-4d7e-be8a-b32d3989de95-utilities\") pod \"community-operators-l2pvb\" (UID: \"e89f69df-9d96-4d7e-be8a-b32d3989de95\") " pod="openshift-marketplace/community-operators-l2pvb" Dec 10 19:35:09 crc kubenswrapper[4828]: I1210 19:35:09.033456 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e89f69df-9d96-4d7e-be8a-b32d3989de95-utilities\") pod \"community-operators-l2pvb\" (UID: \"e89f69df-9d96-4d7e-be8a-b32d3989de95\") " pod="openshift-marketplace/community-operators-l2pvb" Dec 10 19:35:09 crc kubenswrapper[4828]: I1210 19:35:09.033661 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e89f69df-9d96-4d7e-be8a-b32d3989de95-catalog-content\") pod \"community-operators-l2pvb\" (UID: \"e89f69df-9d96-4d7e-be8a-b32d3989de95\") " pod="openshift-marketplace/community-operators-l2pvb" Dec 10 19:35:09 crc kubenswrapper[4828]: I1210 19:35:09.059477 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jswb4\" (UniqueName: \"kubernetes.io/projected/e89f69df-9d96-4d7e-be8a-b32d3989de95-kube-api-access-jswb4\") pod \"community-operators-l2pvb\" (UID: \"e89f69df-9d96-4d7e-be8a-b32d3989de95\") " pod="openshift-marketplace/community-operators-l2pvb" Dec 10 19:35:09 crc kubenswrapper[4828]: I1210 19:35:09.102482 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l2pvb" Dec 10 19:35:09 crc kubenswrapper[4828]: I1210 19:35:09.649598 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l2pvb"] Dec 10 19:35:10 crc kubenswrapper[4828]: I1210 19:35:10.144453 4828 generic.go:334] "Generic (PLEG): container finished" podID="e89f69df-9d96-4d7e-be8a-b32d3989de95" containerID="db5874391b97f553ab846087b8e5851157a441ed7d82fe5cb0f9ce2c49b5484d" exitCode=0 Dec 10 19:35:10 crc kubenswrapper[4828]: I1210 19:35:10.144513 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2pvb" event={"ID":"e89f69df-9d96-4d7e-be8a-b32d3989de95","Type":"ContainerDied","Data":"db5874391b97f553ab846087b8e5851157a441ed7d82fe5cb0f9ce2c49b5484d"} Dec 10 19:35:10 crc kubenswrapper[4828]: I1210 19:35:10.146309 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2pvb" event={"ID":"e89f69df-9d96-4d7e-be8a-b32d3989de95","Type":"ContainerStarted","Data":"ccceaaad2b5fff44f3c222125cba8b3aef82209a0cfc0b6e8030ce2f3536569c"} Dec 10 19:35:12 crc kubenswrapper[4828]: I1210 19:35:12.166531 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2pvb" event={"ID":"e89f69df-9d96-4d7e-be8a-b32d3989de95","Type":"ContainerStarted","Data":"4336dbb5cc73b24a09f274a5612e44d3de19bcf388d3bfade66a1245f81fbcd1"} Dec 10 19:35:13 crc kubenswrapper[4828]: I1210 19:35:13.179225 4828 generic.go:334] "Generic (PLEG): container finished" podID="e89f69df-9d96-4d7e-be8a-b32d3989de95" containerID="4336dbb5cc73b24a09f274a5612e44d3de19bcf388d3bfade66a1245f81fbcd1" exitCode=0 Dec 10 19:35:13 crc kubenswrapper[4828]: I1210 19:35:13.179303 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2pvb" event={"ID":"e89f69df-9d96-4d7e-be8a-b32d3989de95","Type":"ContainerDied","Data":"4336dbb5cc73b24a09f274a5612e44d3de19bcf388d3bfade66a1245f81fbcd1"} Dec 10 19:35:14 crc kubenswrapper[4828]: I1210 19:35:14.203549 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2pvb" event={"ID":"e89f69df-9d96-4d7e-be8a-b32d3989de95","Type":"ContainerStarted","Data":"6155aa2b072c5981c9166813f00370772a0a4643ba7ce734ea2f719719700ed8"} Dec 10 19:35:14 crc kubenswrapper[4828]: I1210 19:35:14.270264 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l2pvb" podStartSLOduration=2.660774038 podStartE2EDuration="6.270242932s" podCreationTimestamp="2025-12-10 19:35:08 +0000 UTC" firstStartedPulling="2025-12-10 19:35:10.146200957 +0000 UTC m=+2390.656811962" lastFinishedPulling="2025-12-10 19:35:13.755669851 +0000 UTC m=+2394.266280856" observedRunningTime="2025-12-10 19:35:14.242200175 +0000 UTC m=+2394.752811180" watchObservedRunningTime="2025-12-10 19:35:14.270242932 +0000 UTC m=+2394.780853937" Dec 10 19:35:19 crc kubenswrapper[4828]: I1210 19:35:19.102741 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l2pvb" Dec 10 19:35:19 crc kubenswrapper[4828]: I1210 19:35:19.103336 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l2pvb" Dec 10 19:35:19 crc kubenswrapper[4828]: I1210 19:35:19.183117 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l2pvb" Dec 10 19:35:19 crc kubenswrapper[4828]: I1210 19:35:19.321680 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l2pvb" Dec 10 19:35:19 crc kubenswrapper[4828]: I1210 19:35:19.435038 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l2pvb"] Dec 10 19:35:21 crc kubenswrapper[4828]: I1210 19:35:21.273255 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l2pvb" podUID="e89f69df-9d96-4d7e-be8a-b32d3989de95" containerName="registry-server" containerID="cri-o://6155aa2b072c5981c9166813f00370772a0a4643ba7ce734ea2f719719700ed8" gracePeriod=2 Dec 10 19:35:21 crc kubenswrapper[4828]: I1210 19:35:21.786346 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l2pvb" Dec 10 19:35:21 crc kubenswrapper[4828]: I1210 19:35:21.789062 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:35:21 crc kubenswrapper[4828]: E1210 19:35:21.789441 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:35:21 crc kubenswrapper[4828]: I1210 19:35:21.944959 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e89f69df-9d96-4d7e-be8a-b32d3989de95-utilities\") pod \"e89f69df-9d96-4d7e-be8a-b32d3989de95\" (UID: \"e89f69df-9d96-4d7e-be8a-b32d3989de95\") " Dec 10 19:35:21 crc kubenswrapper[4828]: I1210 19:35:21.945039 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jswb4\" (UniqueName: \"kubernetes.io/projected/e89f69df-9d96-4d7e-be8a-b32d3989de95-kube-api-access-jswb4\") pod \"e89f69df-9d96-4d7e-be8a-b32d3989de95\" (UID: \"e89f69df-9d96-4d7e-be8a-b32d3989de95\") " Dec 10 19:35:21 crc kubenswrapper[4828]: I1210 19:35:21.945066 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e89f69df-9d96-4d7e-be8a-b32d3989de95-catalog-content\") pod \"e89f69df-9d96-4d7e-be8a-b32d3989de95\" (UID: \"e89f69df-9d96-4d7e-be8a-b32d3989de95\") " Dec 10 19:35:21 crc kubenswrapper[4828]: I1210 19:35:21.946241 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e89f69df-9d96-4d7e-be8a-b32d3989de95-utilities" (OuterVolumeSpecName: "utilities") pod "e89f69df-9d96-4d7e-be8a-b32d3989de95" (UID: "e89f69df-9d96-4d7e-be8a-b32d3989de95"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:35:21 crc kubenswrapper[4828]: I1210 19:35:21.951246 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e89f69df-9d96-4d7e-be8a-b32d3989de95-kube-api-access-jswb4" (OuterVolumeSpecName: "kube-api-access-jswb4") pod "e89f69df-9d96-4d7e-be8a-b32d3989de95" (UID: "e89f69df-9d96-4d7e-be8a-b32d3989de95"). InnerVolumeSpecName "kube-api-access-jswb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:35:21 crc kubenswrapper[4828]: I1210 19:35:21.999174 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e89f69df-9d96-4d7e-be8a-b32d3989de95-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e89f69df-9d96-4d7e-be8a-b32d3989de95" (UID: "e89f69df-9d96-4d7e-be8a-b32d3989de95"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.048336 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e89f69df-9d96-4d7e-be8a-b32d3989de95-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.048370 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jswb4\" (UniqueName: \"kubernetes.io/projected/e89f69df-9d96-4d7e-be8a-b32d3989de95-kube-api-access-jswb4\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.048380 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e89f69df-9d96-4d7e-be8a-b32d3989de95-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.293529 4828 generic.go:334] "Generic (PLEG): container finished" podID="e89f69df-9d96-4d7e-be8a-b32d3989de95" containerID="6155aa2b072c5981c9166813f00370772a0a4643ba7ce734ea2f719719700ed8" exitCode=0 Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.293576 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2pvb" event={"ID":"e89f69df-9d96-4d7e-be8a-b32d3989de95","Type":"ContainerDied","Data":"6155aa2b072c5981c9166813f00370772a0a4643ba7ce734ea2f719719700ed8"} Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.293813 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l2pvb" event={"ID":"e89f69df-9d96-4d7e-be8a-b32d3989de95","Type":"ContainerDied","Data":"ccceaaad2b5fff44f3c222125cba8b3aef82209a0cfc0b6e8030ce2f3536569c"} Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.293838 4828 scope.go:117] "RemoveContainer" containerID="6155aa2b072c5981c9166813f00370772a0a4643ba7ce734ea2f719719700ed8" Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.293645 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l2pvb" Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.331923 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l2pvb"] Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.337240 4828 scope.go:117] "RemoveContainer" containerID="4336dbb5cc73b24a09f274a5612e44d3de19bcf388d3bfade66a1245f81fbcd1" Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.343431 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l2pvb"] Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.364678 4828 scope.go:117] "RemoveContainer" containerID="db5874391b97f553ab846087b8e5851157a441ed7d82fe5cb0f9ce2c49b5484d" Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.414698 4828 scope.go:117] "RemoveContainer" containerID="6155aa2b072c5981c9166813f00370772a0a4643ba7ce734ea2f719719700ed8" Dec 10 19:35:22 crc kubenswrapper[4828]: E1210 19:35:22.415165 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6155aa2b072c5981c9166813f00370772a0a4643ba7ce734ea2f719719700ed8\": container with ID starting with 6155aa2b072c5981c9166813f00370772a0a4643ba7ce734ea2f719719700ed8 not found: ID does not exist" containerID="6155aa2b072c5981c9166813f00370772a0a4643ba7ce734ea2f719719700ed8" Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.415200 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6155aa2b072c5981c9166813f00370772a0a4643ba7ce734ea2f719719700ed8"} err="failed to get container status \"6155aa2b072c5981c9166813f00370772a0a4643ba7ce734ea2f719719700ed8\": rpc error: code = NotFound desc = could not find container \"6155aa2b072c5981c9166813f00370772a0a4643ba7ce734ea2f719719700ed8\": container with ID starting with 6155aa2b072c5981c9166813f00370772a0a4643ba7ce734ea2f719719700ed8 not found: ID does not exist" Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.415223 4828 scope.go:117] "RemoveContainer" containerID="4336dbb5cc73b24a09f274a5612e44d3de19bcf388d3bfade66a1245f81fbcd1" Dec 10 19:35:22 crc kubenswrapper[4828]: E1210 19:35:22.415777 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4336dbb5cc73b24a09f274a5612e44d3de19bcf388d3bfade66a1245f81fbcd1\": container with ID starting with 4336dbb5cc73b24a09f274a5612e44d3de19bcf388d3bfade66a1245f81fbcd1 not found: ID does not exist" containerID="4336dbb5cc73b24a09f274a5612e44d3de19bcf388d3bfade66a1245f81fbcd1" Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.415877 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4336dbb5cc73b24a09f274a5612e44d3de19bcf388d3bfade66a1245f81fbcd1"} err="failed to get container status \"4336dbb5cc73b24a09f274a5612e44d3de19bcf388d3bfade66a1245f81fbcd1\": rpc error: code = NotFound desc = could not find container \"4336dbb5cc73b24a09f274a5612e44d3de19bcf388d3bfade66a1245f81fbcd1\": container with ID starting with 4336dbb5cc73b24a09f274a5612e44d3de19bcf388d3bfade66a1245f81fbcd1 not found: ID does not exist" Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.415914 4828 scope.go:117] "RemoveContainer" containerID="db5874391b97f553ab846087b8e5851157a441ed7d82fe5cb0f9ce2c49b5484d" Dec 10 19:35:22 crc kubenswrapper[4828]: E1210 19:35:22.416358 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db5874391b97f553ab846087b8e5851157a441ed7d82fe5cb0f9ce2c49b5484d\": container with ID starting with db5874391b97f553ab846087b8e5851157a441ed7d82fe5cb0f9ce2c49b5484d not found: ID does not exist" containerID="db5874391b97f553ab846087b8e5851157a441ed7d82fe5cb0f9ce2c49b5484d" Dec 10 19:35:22 crc kubenswrapper[4828]: I1210 19:35:22.416387 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db5874391b97f553ab846087b8e5851157a441ed7d82fe5cb0f9ce2c49b5484d"} err="failed to get container status \"db5874391b97f553ab846087b8e5851157a441ed7d82fe5cb0f9ce2c49b5484d\": rpc error: code = NotFound desc = could not find container \"db5874391b97f553ab846087b8e5851157a441ed7d82fe5cb0f9ce2c49b5484d\": container with ID starting with db5874391b97f553ab846087b8e5851157a441ed7d82fe5cb0f9ce2c49b5484d not found: ID does not exist" Dec 10 19:35:23 crc kubenswrapper[4828]: I1210 19:35:23.800117 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e89f69df-9d96-4d7e-be8a-b32d3989de95" path="/var/lib/kubelet/pods/e89f69df-9d96-4d7e-be8a-b32d3989de95/volumes" Dec 10 19:35:25 crc kubenswrapper[4828]: I1210 19:35:25.329088 4828 generic.go:334] "Generic (PLEG): container finished" podID="5726d3d2-c797-4fca-aa6c-8aec98c6061b" containerID="049485cfdbe4e24fe1d11b8c1ee03f62ba5e91996a24354ecc095cf0dc29c76c" exitCode=0 Dec 10 19:35:25 crc kubenswrapper[4828]: I1210 19:35:25.329194 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" event={"ID":"5726d3d2-c797-4fca-aa6c-8aec98c6061b","Type":"ContainerDied","Data":"049485cfdbe4e24fe1d11b8c1ee03f62ba5e91996a24354ecc095cf0dc29c76c"} Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.281971 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.351694 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" event={"ID":"5726d3d2-c797-4fca-aa6c-8aec98c6061b","Type":"ContainerDied","Data":"9d6bfed25e610e855bc94ca90deb47d28aed09a98803a8fba541cf5ac974afbd"} Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.351733 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d6bfed25e610e855bc94ca90deb47d28aed09a98803a8fba541cf5ac974afbd" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.351811 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.375472 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-telemetry-power-monitoring-combined-ca-bundle\") pod \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.375519 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-inventory\") pod \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.375551 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-ovn-default-certs-0\") pod \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.375646 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-bootstrap-combined-ca-bundle\") pod \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.375670 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-ovn-combined-ca-bundle\") pod \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.375700 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-neutron-metadata-combined-ca-bundle\") pod \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.375818 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-libvirt-combined-ca-bundle\") pod \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.375843 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.376513 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-nova-combined-ca-bundle\") pod \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.376583 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.376622 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.376695 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-repo-setup-combined-ca-bundle\") pod \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.376727 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-ssh-key\") pod \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.376771 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c878l\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-kube-api-access-c878l\") pod \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.376846 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-telemetry-combined-ca-bundle\") pod \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.376885 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\" (UID: \"5726d3d2-c797-4fca-aa6c-8aec98c6061b\") " Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.384343 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "5726d3d2-c797-4fca-aa6c-8aec98c6061b" (UID: "5726d3d2-c797-4fca-aa6c-8aec98c6061b"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.385224 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "5726d3d2-c797-4fca-aa6c-8aec98c6061b" (UID: "5726d3d2-c797-4fca-aa6c-8aec98c6061b"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.386083 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "5726d3d2-c797-4fca-aa6c-8aec98c6061b" (UID: "5726d3d2-c797-4fca-aa6c-8aec98c6061b"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.389023 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "5726d3d2-c797-4fca-aa6c-8aec98c6061b" (UID: "5726d3d2-c797-4fca-aa6c-8aec98c6061b"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.389140 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "5726d3d2-c797-4fca-aa6c-8aec98c6061b" (UID: "5726d3d2-c797-4fca-aa6c-8aec98c6061b"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.390004 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "5726d3d2-c797-4fca-aa6c-8aec98c6061b" (UID: "5726d3d2-c797-4fca-aa6c-8aec98c6061b"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.393648 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "5726d3d2-c797-4fca-aa6c-8aec98c6061b" (UID: "5726d3d2-c797-4fca-aa6c-8aec98c6061b"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.393724 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "5726d3d2-c797-4fca-aa6c-8aec98c6061b" (UID: "5726d3d2-c797-4fca-aa6c-8aec98c6061b"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.393755 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "5726d3d2-c797-4fca-aa6c-8aec98c6061b" (UID: "5726d3d2-c797-4fca-aa6c-8aec98c6061b"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.394232 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "5726d3d2-c797-4fca-aa6c-8aec98c6061b" (UID: "5726d3d2-c797-4fca-aa6c-8aec98c6061b"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.403430 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-kube-api-access-c878l" (OuterVolumeSpecName: "kube-api-access-c878l") pod "5726d3d2-c797-4fca-aa6c-8aec98c6061b" (UID: "5726d3d2-c797-4fca-aa6c-8aec98c6061b"). InnerVolumeSpecName "kube-api-access-c878l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.404316 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "5726d3d2-c797-4fca-aa6c-8aec98c6061b" (UID: "5726d3d2-c797-4fca-aa6c-8aec98c6061b"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.404756 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "5726d3d2-c797-4fca-aa6c-8aec98c6061b" (UID: "5726d3d2-c797-4fca-aa6c-8aec98c6061b"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.407186 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "5726d3d2-c797-4fca-aa6c-8aec98c6061b" (UID: "5726d3d2-c797-4fca-aa6c-8aec98c6061b"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.432092 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-inventory" (OuterVolumeSpecName: "inventory") pod "5726d3d2-c797-4fca-aa6c-8aec98c6061b" (UID: "5726d3d2-c797-4fca-aa6c-8aec98c6061b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.432521 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5726d3d2-c797-4fca-aa6c-8aec98c6061b" (UID: "5726d3d2-c797-4fca-aa6c-8aec98c6061b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.460189 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx"] Dec 10 19:35:27 crc kubenswrapper[4828]: E1210 19:35:27.460771 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5726d3d2-c797-4fca-aa6c-8aec98c6061b" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.460790 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="5726d3d2-c797-4fca-aa6c-8aec98c6061b" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 10 19:35:27 crc kubenswrapper[4828]: E1210 19:35:27.460841 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e89f69df-9d96-4d7e-be8a-b32d3989de95" containerName="extract-content" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.460848 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="e89f69df-9d96-4d7e-be8a-b32d3989de95" containerName="extract-content" Dec 10 19:35:27 crc kubenswrapper[4828]: E1210 19:35:27.460865 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e89f69df-9d96-4d7e-be8a-b32d3989de95" containerName="extract-utilities" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.460871 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="e89f69df-9d96-4d7e-be8a-b32d3989de95" containerName="extract-utilities" Dec 10 19:35:27 crc kubenswrapper[4828]: E1210 19:35:27.460885 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e89f69df-9d96-4d7e-be8a-b32d3989de95" containerName="registry-server" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.460891 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="e89f69df-9d96-4d7e-be8a-b32d3989de95" containerName="registry-server" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.461101 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="e89f69df-9d96-4d7e-be8a-b32d3989de95" containerName="registry-server" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.461123 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="5726d3d2-c797-4fca-aa6c-8aec98c6061b" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.462043 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.468201 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.472904 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx"] Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.480510 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.480542 4828 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.480555 4828 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.480566 4828 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.480575 4828 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.480588 4828 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.480596 4828 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.480605 4828 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.480617 4828 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.480629 4828 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.480638 4828 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.480646 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.480655 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c878l\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-kube-api-access-c878l\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.480664 4828 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.480672 4828 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/5726d3d2-c797-4fca-aa6c-8aec98c6061b-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.480681 4828 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726d3d2-c797-4fca-aa6c-8aec98c6061b-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.582291 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-f9knx\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.582490 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj9b5\" (UniqueName: \"kubernetes.io/projected/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-kube-api-access-dj9b5\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-f9knx\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.582676 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-f9knx\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.582835 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-f9knx\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.582904 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-f9knx\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.684746 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-f9knx\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.685002 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-f9knx\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.685204 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj9b5\" (UniqueName: \"kubernetes.io/projected/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-kube-api-access-dj9b5\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-f9knx\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.685384 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-f9knx\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.685591 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-f9knx\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.686503 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-f9knx\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.691306 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-f9knx\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.691332 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-f9knx\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.692360 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-f9knx\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.703293 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj9b5\" (UniqueName: \"kubernetes.io/projected/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-kube-api-access-dj9b5\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-f9knx\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:35:27 crc kubenswrapper[4828]: I1210 19:35:27.853514 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:35:28 crc kubenswrapper[4828]: I1210 19:35:28.450101 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx"] Dec 10 19:35:29 crc kubenswrapper[4828]: I1210 19:35:29.370964 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" event={"ID":"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94","Type":"ContainerStarted","Data":"2e02272dda27ec55647f39722aa57a66bcacb0c46372e4c68dbbba50df15d2f3"} Dec 10 19:35:29 crc kubenswrapper[4828]: I1210 19:35:29.371470 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" event={"ID":"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94","Type":"ContainerStarted","Data":"94584453964e42f4388731d9747e17e0ce5cb91745efa28822636ca00568279a"} Dec 10 19:35:29 crc kubenswrapper[4828]: I1210 19:35:29.396268 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" podStartSLOduration=1.931914082 podStartE2EDuration="2.396248295s" podCreationTimestamp="2025-12-10 19:35:27 +0000 UTC" firstStartedPulling="2025-12-10 19:35:28.456894723 +0000 UTC m=+2408.967505728" lastFinishedPulling="2025-12-10 19:35:28.921228936 +0000 UTC m=+2409.431839941" observedRunningTime="2025-12-10 19:35:29.388281153 +0000 UTC m=+2409.898892158" watchObservedRunningTime="2025-12-10 19:35:29.396248295 +0000 UTC m=+2409.906859300" Dec 10 19:35:31 crc kubenswrapper[4828]: I1210 19:35:31.044573 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-7krk4"] Dec 10 19:35:31 crc kubenswrapper[4828]: I1210 19:35:31.056702 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-7krk4"] Dec 10 19:35:31 crc kubenswrapper[4828]: I1210 19:35:31.801476 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="788b1631-a897-4929-972f-4c8b9bf008db" path="/var/lib/kubelet/pods/788b1631-a897-4929-972f-4c8b9bf008db/volumes" Dec 10 19:35:35 crc kubenswrapper[4828]: I1210 19:35:35.789402 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:35:35 crc kubenswrapper[4828]: E1210 19:35:35.790200 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:35:41 crc kubenswrapper[4828]: I1210 19:35:41.750148 4828 scope.go:117] "RemoveContainer" containerID="2ac69f9a7894f1dd0075acf25f01223ed2fcb8080f0630897935c14a98d127ab" Dec 10 19:35:41 crc kubenswrapper[4828]: I1210 19:35:41.801014 4828 scope.go:117] "RemoveContainer" containerID="83c59f81f6ea289aaabe02c0e97b3bb1998fcf5e568319c373edf91d5ae2c18b" Dec 10 19:35:50 crc kubenswrapper[4828]: I1210 19:35:50.788832 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:35:50 crc kubenswrapper[4828]: E1210 19:35:50.790481 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:36:01 crc kubenswrapper[4828]: I1210 19:36:01.789453 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:36:01 crc kubenswrapper[4828]: E1210 19:36:01.790232 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:36:16 crc kubenswrapper[4828]: I1210 19:36:16.789868 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:36:16 crc kubenswrapper[4828]: E1210 19:36:16.790534 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:36:29 crc kubenswrapper[4828]: I1210 19:36:29.789578 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:36:29 crc kubenswrapper[4828]: E1210 19:36:29.790502 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:36:32 crc kubenswrapper[4828]: I1210 19:36:32.074207 4828 generic.go:334] "Generic (PLEG): container finished" podID="6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94" containerID="2e02272dda27ec55647f39722aa57a66bcacb0c46372e4c68dbbba50df15d2f3" exitCode=0 Dec 10 19:36:32 crc kubenswrapper[4828]: I1210 19:36:32.074323 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" event={"ID":"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94","Type":"ContainerDied","Data":"2e02272dda27ec55647f39722aa57a66bcacb0c46372e4c68dbbba50df15d2f3"} Dec 10 19:36:33 crc kubenswrapper[4828]: I1210 19:36:33.510516 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:36:33 crc kubenswrapper[4828]: I1210 19:36:33.554513 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-inventory\") pod \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " Dec 10 19:36:33 crc kubenswrapper[4828]: I1210 19:36:33.554692 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj9b5\" (UniqueName: \"kubernetes.io/projected/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-kube-api-access-dj9b5\") pod \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " Dec 10 19:36:33 crc kubenswrapper[4828]: I1210 19:36:33.554840 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ovn-combined-ca-bundle\") pod \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " Dec 10 19:36:33 crc kubenswrapper[4828]: I1210 19:36:33.554877 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ssh-key\") pod \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " Dec 10 19:36:33 crc kubenswrapper[4828]: I1210 19:36:33.554966 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ovncontroller-config-0\") pod \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\" (UID: \"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94\") " Dec 10 19:36:33 crc kubenswrapper[4828]: I1210 19:36:33.563909 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94" (UID: "6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:36:33 crc kubenswrapper[4828]: I1210 19:36:33.565845 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-kube-api-access-dj9b5" (OuterVolumeSpecName: "kube-api-access-dj9b5") pod "6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94" (UID: "6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94"). InnerVolumeSpecName "kube-api-access-dj9b5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:36:33 crc kubenswrapper[4828]: I1210 19:36:33.628531 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94" (UID: "6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:36:33 crc kubenswrapper[4828]: I1210 19:36:33.630913 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94" (UID: "6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:36:33 crc kubenswrapper[4828]: I1210 19:36:33.632028 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-inventory" (OuterVolumeSpecName: "inventory") pod "6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94" (UID: "6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:36:33 crc kubenswrapper[4828]: I1210 19:36:33.658390 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:36:33 crc kubenswrapper[4828]: I1210 19:36:33.658422 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj9b5\" (UniqueName: \"kubernetes.io/projected/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-kube-api-access-dj9b5\") on node \"crc\" DevicePath \"\"" Dec 10 19:36:33 crc kubenswrapper[4828]: I1210 19:36:33.658438 4828 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:36:33 crc kubenswrapper[4828]: I1210 19:36:33.658450 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:36:33 crc kubenswrapper[4828]: I1210 19:36:33.658460 4828 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.101084 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" event={"ID":"6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94","Type":"ContainerDied","Data":"94584453964e42f4388731d9747e17e0ce5cb91745efa28822636ca00568279a"} Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.101122 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94584453964e42f4388731d9747e17e0ce5cb91745efa28822636ca00568279a" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.101236 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-f9knx" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.232219 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj"] Dec 10 19:36:34 crc kubenswrapper[4828]: E1210 19:36:34.232778 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.232840 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.233127 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.234082 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.236149 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.236388 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.237710 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.237926 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.238098 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.238248 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.241274 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj"] Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.277834 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.278404 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.278448 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.278505 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.278552 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.278614 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fczqr\" (UniqueName: \"kubernetes.io/projected/c502be9a-c873-4b9b-97e8-e6461052d252-kube-api-access-fczqr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.380505 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.380582 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fczqr\" (UniqueName: \"kubernetes.io/projected/c502be9a-c873-4b9b-97e8-e6461052d252-kube-api-access-fczqr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.380620 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.380765 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.380787 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.380840 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.387520 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.387917 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.388663 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.388725 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.388977 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.396894 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fczqr\" (UniqueName: \"kubernetes.io/projected/c502be9a-c873-4b9b-97e8-e6461052d252-kube-api-access-fczqr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:34 crc kubenswrapper[4828]: I1210 19:36:34.558568 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:36:35 crc kubenswrapper[4828]: I1210 19:36:35.130232 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj"] Dec 10 19:36:36 crc kubenswrapper[4828]: I1210 19:36:36.121198 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" event={"ID":"c502be9a-c873-4b9b-97e8-e6461052d252","Type":"ContainerStarted","Data":"cf422b9b56cd0e6ddf464378666886e964d460a6e52e0f2805054a611f75d917"} Dec 10 19:36:36 crc kubenswrapper[4828]: I1210 19:36:36.122711 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" event={"ID":"c502be9a-c873-4b9b-97e8-e6461052d252","Type":"ContainerStarted","Data":"74903834ead0812f60b9ea492a034026fd1cfc0ee7bb0764a79b6d02dac883f9"} Dec 10 19:36:36 crc kubenswrapper[4828]: I1210 19:36:36.145043 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" podStartSLOduration=1.686103533 podStartE2EDuration="2.145011583s" podCreationTimestamp="2025-12-10 19:36:34 +0000 UTC" firstStartedPulling="2025-12-10 19:36:35.133333486 +0000 UTC m=+2475.643944491" lastFinishedPulling="2025-12-10 19:36:35.592241536 +0000 UTC m=+2476.102852541" observedRunningTime="2025-12-10 19:36:36.13888764 +0000 UTC m=+2476.649498645" watchObservedRunningTime="2025-12-10 19:36:36.145011583 +0000 UTC m=+2476.655622628" Dec 10 19:36:40 crc kubenswrapper[4828]: I1210 19:36:40.788727 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:36:40 crc kubenswrapper[4828]: E1210 19:36:40.789551 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:36:53 crc kubenswrapper[4828]: I1210 19:36:53.788815 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:36:53 crc kubenswrapper[4828]: E1210 19:36:53.789593 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:37:07 crc kubenswrapper[4828]: I1210 19:37:07.790140 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:37:07 crc kubenswrapper[4828]: E1210 19:37:07.791475 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:37:18 crc kubenswrapper[4828]: I1210 19:37:18.789782 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:37:18 crc kubenswrapper[4828]: E1210 19:37:18.790842 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:37:22 crc kubenswrapper[4828]: I1210 19:37:22.623310 4828 generic.go:334] "Generic (PLEG): container finished" podID="c502be9a-c873-4b9b-97e8-e6461052d252" containerID="cf422b9b56cd0e6ddf464378666886e964d460a6e52e0f2805054a611f75d917" exitCode=0 Dec 10 19:37:22 crc kubenswrapper[4828]: I1210 19:37:22.623399 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" event={"ID":"c502be9a-c873-4b9b-97e8-e6461052d252","Type":"ContainerDied","Data":"cf422b9b56cd0e6ddf464378666886e964d460a6e52e0f2805054a611f75d917"} Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.088994 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.142562 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-nova-metadata-neutron-config-0\") pod \"c502be9a-c873-4b9b-97e8-e6461052d252\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.142772 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-inventory\") pod \"c502be9a-c873-4b9b-97e8-e6461052d252\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.142865 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-ssh-key\") pod \"c502be9a-c873-4b9b-97e8-e6461052d252\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.142897 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-neutron-metadata-combined-ca-bundle\") pod \"c502be9a-c873-4b9b-97e8-e6461052d252\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.142934 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fczqr\" (UniqueName: \"kubernetes.io/projected/c502be9a-c873-4b9b-97e8-e6461052d252-kube-api-access-fczqr\") pod \"c502be9a-c873-4b9b-97e8-e6461052d252\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.142985 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-neutron-ovn-metadata-agent-neutron-config-0\") pod \"c502be9a-c873-4b9b-97e8-e6461052d252\" (UID: \"c502be9a-c873-4b9b-97e8-e6461052d252\") " Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.150189 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c502be9a-c873-4b9b-97e8-e6461052d252-kube-api-access-fczqr" (OuterVolumeSpecName: "kube-api-access-fczqr") pod "c502be9a-c873-4b9b-97e8-e6461052d252" (UID: "c502be9a-c873-4b9b-97e8-e6461052d252"). InnerVolumeSpecName "kube-api-access-fczqr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.164130 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "c502be9a-c873-4b9b-97e8-e6461052d252" (UID: "c502be9a-c873-4b9b-97e8-e6461052d252"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.179030 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-inventory" (OuterVolumeSpecName: "inventory") pod "c502be9a-c873-4b9b-97e8-e6461052d252" (UID: "c502be9a-c873-4b9b-97e8-e6461052d252"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.186633 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "c502be9a-c873-4b9b-97e8-e6461052d252" (UID: "c502be9a-c873-4b9b-97e8-e6461052d252"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.188001 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c502be9a-c873-4b9b-97e8-e6461052d252" (UID: "c502be9a-c873-4b9b-97e8-e6461052d252"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.188732 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "c502be9a-c873-4b9b-97e8-e6461052d252" (UID: "c502be9a-c873-4b9b-97e8-e6461052d252"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.245852 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.245890 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.245902 4828 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.245917 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fczqr\" (UniqueName: \"kubernetes.io/projected/c502be9a-c873-4b9b-97e8-e6461052d252-kube-api-access-fczqr\") on node \"crc\" DevicePath \"\"" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.245928 4828 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.245939 4828 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c502be9a-c873-4b9b-97e8-e6461052d252-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.653221 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" event={"ID":"c502be9a-c873-4b9b-97e8-e6461052d252","Type":"ContainerDied","Data":"74903834ead0812f60b9ea492a034026fd1cfc0ee7bb0764a79b6d02dac883f9"} Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.653542 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74903834ead0812f60b9ea492a034026fd1cfc0ee7bb0764a79b6d02dac883f9" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.653339 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.751779 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w"] Dec 10 19:37:24 crc kubenswrapper[4828]: E1210 19:37:24.752332 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c502be9a-c873-4b9b-97e8-e6461052d252" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.752358 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c502be9a-c873-4b9b-97e8-e6461052d252" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.752671 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c502be9a-c873-4b9b-97e8-e6461052d252" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.753701 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.755508 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.755973 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.755979 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.757163 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.758469 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.770555 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w"] Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.860468 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.861594 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlbsz\" (UniqueName: \"kubernetes.io/projected/791b45cd-a74c-493b-b33b-e97c642318dc-kube-api-access-xlbsz\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.861626 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.861695 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.861762 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.963660 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.963812 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlbsz\" (UniqueName: \"kubernetes.io/projected/791b45cd-a74c-493b-b33b-e97c642318dc-kube-api-access-xlbsz\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.963844 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.963912 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.963972 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.968691 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.969421 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.972138 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.976634 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:37:24 crc kubenswrapper[4828]: I1210 19:37:24.981409 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlbsz\" (UniqueName: \"kubernetes.io/projected/791b45cd-a74c-493b-b33b-e97c642318dc-kube-api-access-xlbsz\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:37:25 crc kubenswrapper[4828]: I1210 19:37:25.079095 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:37:25 crc kubenswrapper[4828]: I1210 19:37:25.650631 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w"] Dec 10 19:37:25 crc kubenswrapper[4828]: I1210 19:37:25.668175 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" event={"ID":"791b45cd-a74c-493b-b33b-e97c642318dc","Type":"ContainerStarted","Data":"f7d14714c82d3c8dcb5d17030eee9e911035aa187d4263d566cd9fb7b8cf95af"} Dec 10 19:37:26 crc kubenswrapper[4828]: I1210 19:37:26.679367 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" event={"ID":"791b45cd-a74c-493b-b33b-e97c642318dc","Type":"ContainerStarted","Data":"8454e266c8f2c0dd14e01f0ad9e5f66acf71417c617bad9dc5663fe1366eae29"} Dec 10 19:37:26 crc kubenswrapper[4828]: I1210 19:37:26.709256 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" podStartSLOduration=2.137977243 podStartE2EDuration="2.709233523s" podCreationTimestamp="2025-12-10 19:37:24 +0000 UTC" firstStartedPulling="2025-12-10 19:37:25.653313048 +0000 UTC m=+2526.163924053" lastFinishedPulling="2025-12-10 19:37:26.224569328 +0000 UTC m=+2526.735180333" observedRunningTime="2025-12-10 19:37:26.692726974 +0000 UTC m=+2527.203337979" watchObservedRunningTime="2025-12-10 19:37:26.709233523 +0000 UTC m=+2527.219844538" Dec 10 19:37:32 crc kubenswrapper[4828]: I1210 19:37:32.789141 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:37:33 crc kubenswrapper[4828]: I1210 19:37:33.748488 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"823c86952f20d947f08d0d8bb359ea62a6adbee8aa27350f3f768ba427671ebc"} Dec 10 19:39:51 crc kubenswrapper[4828]: I1210 19:39:51.230529 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:39:51 crc kubenswrapper[4828]: I1210 19:39:51.231384 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:40:21 crc kubenswrapper[4828]: I1210 19:40:21.230827 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:40:21 crc kubenswrapper[4828]: I1210 19:40:21.231351 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:40:51 crc kubenswrapper[4828]: I1210 19:40:51.230221 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:40:51 crc kubenswrapper[4828]: I1210 19:40:51.230856 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:40:51 crc kubenswrapper[4828]: I1210 19:40:51.230913 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 19:40:51 crc kubenswrapper[4828]: I1210 19:40:51.233519 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"823c86952f20d947f08d0d8bb359ea62a6adbee8aa27350f3f768ba427671ebc"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:40:51 crc kubenswrapper[4828]: I1210 19:40:51.233604 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://823c86952f20d947f08d0d8bb359ea62a6adbee8aa27350f3f768ba427671ebc" gracePeriod=600 Dec 10 19:40:51 crc kubenswrapper[4828]: I1210 19:40:51.962218 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="823c86952f20d947f08d0d8bb359ea62a6adbee8aa27350f3f768ba427671ebc" exitCode=0 Dec 10 19:40:51 crc kubenswrapper[4828]: I1210 19:40:51.962781 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"823c86952f20d947f08d0d8bb359ea62a6adbee8aa27350f3f768ba427671ebc"} Dec 10 19:40:51 crc kubenswrapper[4828]: I1210 19:40:51.962835 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83"} Dec 10 19:40:51 crc kubenswrapper[4828]: I1210 19:40:51.962856 4828 scope.go:117] "RemoveContainer" containerID="5bc114f76e762981c9c50294db46df5e7c52db4b7fd645b177af7dad4d17a3a6" Dec 10 19:41:45 crc kubenswrapper[4828]: I1210 19:41:45.582140 4828 generic.go:334] "Generic (PLEG): container finished" podID="791b45cd-a74c-493b-b33b-e97c642318dc" containerID="8454e266c8f2c0dd14e01f0ad9e5f66acf71417c617bad9dc5663fe1366eae29" exitCode=0 Dec 10 19:41:45 crc kubenswrapper[4828]: I1210 19:41:45.582223 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" event={"ID":"791b45cd-a74c-493b-b33b-e97c642318dc","Type":"ContainerDied","Data":"8454e266c8f2c0dd14e01f0ad9e5f66acf71417c617bad9dc5663fe1366eae29"} Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.110228 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.193065 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-libvirt-secret-0\") pod \"791b45cd-a74c-493b-b33b-e97c642318dc\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.193385 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-ssh-key\") pod \"791b45cd-a74c-493b-b33b-e97c642318dc\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.193460 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-libvirt-combined-ca-bundle\") pod \"791b45cd-a74c-493b-b33b-e97c642318dc\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.193648 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-inventory\") pod \"791b45cd-a74c-493b-b33b-e97c642318dc\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.193721 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlbsz\" (UniqueName: \"kubernetes.io/projected/791b45cd-a74c-493b-b33b-e97c642318dc-kube-api-access-xlbsz\") pod \"791b45cd-a74c-493b-b33b-e97c642318dc\" (UID: \"791b45cd-a74c-493b-b33b-e97c642318dc\") " Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.199766 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/791b45cd-a74c-493b-b33b-e97c642318dc-kube-api-access-xlbsz" (OuterVolumeSpecName: "kube-api-access-xlbsz") pod "791b45cd-a74c-493b-b33b-e97c642318dc" (UID: "791b45cd-a74c-493b-b33b-e97c642318dc"). InnerVolumeSpecName "kube-api-access-xlbsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.206275 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "791b45cd-a74c-493b-b33b-e97c642318dc" (UID: "791b45cd-a74c-493b-b33b-e97c642318dc"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.233742 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "791b45cd-a74c-493b-b33b-e97c642318dc" (UID: "791b45cd-a74c-493b-b33b-e97c642318dc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.237940 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "791b45cd-a74c-493b-b33b-e97c642318dc" (UID: "791b45cd-a74c-493b-b33b-e97c642318dc"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.241055 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-inventory" (OuterVolumeSpecName: "inventory") pod "791b45cd-a74c-493b-b33b-e97c642318dc" (UID: "791b45cd-a74c-493b-b33b-e97c642318dc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.295843 4828 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.295879 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.295893 4828 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.295907 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/791b45cd-a74c-493b-b33b-e97c642318dc-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.295917 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlbsz\" (UniqueName: \"kubernetes.io/projected/791b45cd-a74c-493b-b33b-e97c642318dc-kube-api-access-xlbsz\") on node \"crc\" DevicePath \"\"" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.603214 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" event={"ID":"791b45cd-a74c-493b-b33b-e97c642318dc","Type":"ContainerDied","Data":"f7d14714c82d3c8dcb5d17030eee9e911035aa187d4263d566cd9fb7b8cf95af"} Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.603259 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7d14714c82d3c8dcb5d17030eee9e911035aa187d4263d566cd9fb7b8cf95af" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.603285 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.697560 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn"] Dec 10 19:41:47 crc kubenswrapper[4828]: E1210 19:41:47.698066 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="791b45cd-a74c-493b-b33b-e97c642318dc" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.698084 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="791b45cd-a74c-493b-b33b-e97c642318dc" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.698309 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="791b45cd-a74c-493b-b33b-e97c642318dc" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.699184 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.701720 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.702015 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.702337 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.702381 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.705299 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.706667 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.706863 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.708978 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn"] Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.806907 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.806951 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.807016 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.807511 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgwf5\" (UniqueName: \"kubernetes.io/projected/7dc2751a-2c10-409f-8313-f6e78fb87ba3-kube-api-access-pgwf5\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.807600 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.807631 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.807660 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.807692 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.807849 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.910878 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.911305 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.911390 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.911452 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgwf5\" (UniqueName: \"kubernetes.io/projected/7dc2751a-2c10-409f-8313-f6e78fb87ba3-kube-api-access-pgwf5\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.911520 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.912551 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.911541 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.912658 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.912689 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.913322 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.916730 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.917091 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.917604 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.918033 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.918298 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.918496 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.920836 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:47 crc kubenswrapper[4828]: I1210 19:41:47.928260 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgwf5\" (UniqueName: \"kubernetes.io/projected/7dc2751a-2c10-409f-8313-f6e78fb87ba3-kube-api-access-pgwf5\") pod \"nova-edpm-deployment-openstack-edpm-ipam-475xn\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:48 crc kubenswrapper[4828]: I1210 19:41:48.014997 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:41:48 crc kubenswrapper[4828]: I1210 19:41:48.572147 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn"] Dec 10 19:41:48 crc kubenswrapper[4828]: I1210 19:41:48.575208 4828 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:41:48 crc kubenswrapper[4828]: I1210 19:41:48.615517 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" event={"ID":"7dc2751a-2c10-409f-8313-f6e78fb87ba3","Type":"ContainerStarted","Data":"d2d9ac3b54116080316cf2edaf5f90061b13675385d5f55968504e7aa65184ee"} Dec 10 19:41:49 crc kubenswrapper[4828]: I1210 19:41:49.628148 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" event={"ID":"7dc2751a-2c10-409f-8313-f6e78fb87ba3","Type":"ContainerStarted","Data":"0352a738064fef72110c449470c67b89f3d647ef0fbcf33d751c501457d0b17c"} Dec 10 19:41:49 crc kubenswrapper[4828]: I1210 19:41:49.652203 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" podStartSLOduration=2.027774123 podStartE2EDuration="2.652185066s" podCreationTimestamp="2025-12-10 19:41:47 +0000 UTC" firstStartedPulling="2025-12-10 19:41:48.574941664 +0000 UTC m=+2789.085552679" lastFinishedPulling="2025-12-10 19:41:49.199352607 +0000 UTC m=+2789.709963622" observedRunningTime="2025-12-10 19:41:49.646838914 +0000 UTC m=+2790.157449939" watchObservedRunningTime="2025-12-10 19:41:49.652185066 +0000 UTC m=+2790.162796061" Dec 10 19:42:09 crc kubenswrapper[4828]: I1210 19:42:09.704586 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9jxhs"] Dec 10 19:42:09 crc kubenswrapper[4828]: I1210 19:42:09.707937 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9jxhs" Dec 10 19:42:09 crc kubenswrapper[4828]: I1210 19:42:09.720864 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9jxhs"] Dec 10 19:42:09 crc kubenswrapper[4828]: I1210 19:42:09.852019 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtj48\" (UniqueName: \"kubernetes.io/projected/ead62f92-39c7-46bb-be5d-f721322ee339-kube-api-access-rtj48\") pod \"redhat-operators-9jxhs\" (UID: \"ead62f92-39c7-46bb-be5d-f721322ee339\") " pod="openshift-marketplace/redhat-operators-9jxhs" Dec 10 19:42:09 crc kubenswrapper[4828]: I1210 19:42:09.852080 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ead62f92-39c7-46bb-be5d-f721322ee339-utilities\") pod \"redhat-operators-9jxhs\" (UID: \"ead62f92-39c7-46bb-be5d-f721322ee339\") " pod="openshift-marketplace/redhat-operators-9jxhs" Dec 10 19:42:09 crc kubenswrapper[4828]: I1210 19:42:09.852156 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ead62f92-39c7-46bb-be5d-f721322ee339-catalog-content\") pod \"redhat-operators-9jxhs\" (UID: \"ead62f92-39c7-46bb-be5d-f721322ee339\") " pod="openshift-marketplace/redhat-operators-9jxhs" Dec 10 19:42:09 crc kubenswrapper[4828]: I1210 19:42:09.955330 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtj48\" (UniqueName: \"kubernetes.io/projected/ead62f92-39c7-46bb-be5d-f721322ee339-kube-api-access-rtj48\") pod \"redhat-operators-9jxhs\" (UID: \"ead62f92-39c7-46bb-be5d-f721322ee339\") " pod="openshift-marketplace/redhat-operators-9jxhs" Dec 10 19:42:09 crc kubenswrapper[4828]: I1210 19:42:09.955392 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ead62f92-39c7-46bb-be5d-f721322ee339-utilities\") pod \"redhat-operators-9jxhs\" (UID: \"ead62f92-39c7-46bb-be5d-f721322ee339\") " pod="openshift-marketplace/redhat-operators-9jxhs" Dec 10 19:42:09 crc kubenswrapper[4828]: I1210 19:42:09.955448 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ead62f92-39c7-46bb-be5d-f721322ee339-catalog-content\") pod \"redhat-operators-9jxhs\" (UID: \"ead62f92-39c7-46bb-be5d-f721322ee339\") " pod="openshift-marketplace/redhat-operators-9jxhs" Dec 10 19:42:09 crc kubenswrapper[4828]: I1210 19:42:09.956195 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ead62f92-39c7-46bb-be5d-f721322ee339-utilities\") pod \"redhat-operators-9jxhs\" (UID: \"ead62f92-39c7-46bb-be5d-f721322ee339\") " pod="openshift-marketplace/redhat-operators-9jxhs" Dec 10 19:42:09 crc kubenswrapper[4828]: I1210 19:42:09.956243 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ead62f92-39c7-46bb-be5d-f721322ee339-catalog-content\") pod \"redhat-operators-9jxhs\" (UID: \"ead62f92-39c7-46bb-be5d-f721322ee339\") " pod="openshift-marketplace/redhat-operators-9jxhs" Dec 10 19:42:09 crc kubenswrapper[4828]: I1210 19:42:09.978292 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtj48\" (UniqueName: \"kubernetes.io/projected/ead62f92-39c7-46bb-be5d-f721322ee339-kube-api-access-rtj48\") pod \"redhat-operators-9jxhs\" (UID: \"ead62f92-39c7-46bb-be5d-f721322ee339\") " pod="openshift-marketplace/redhat-operators-9jxhs" Dec 10 19:42:10 crc kubenswrapper[4828]: I1210 19:42:10.033004 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9jxhs" Dec 10 19:42:10 crc kubenswrapper[4828]: W1210 19:42:10.486990 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podead62f92_39c7_46bb_be5d_f721322ee339.slice/crio-21c0f723a88cc3b0ca87807477d80d8bc59c311630a3634942e1310bb0f7fb99 WatchSource:0}: Error finding container 21c0f723a88cc3b0ca87807477d80d8bc59c311630a3634942e1310bb0f7fb99: Status 404 returned error can't find the container with id 21c0f723a88cc3b0ca87807477d80d8bc59c311630a3634942e1310bb0f7fb99 Dec 10 19:42:10 crc kubenswrapper[4828]: I1210 19:42:10.494741 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9jxhs"] Dec 10 19:42:10 crc kubenswrapper[4828]: I1210 19:42:10.855249 4828 generic.go:334] "Generic (PLEG): container finished" podID="ead62f92-39c7-46bb-be5d-f721322ee339" containerID="b5081e46e877e4b88a775bf140bb153b368c60cbb91bf6e9c14ef3be4643017e" exitCode=0 Dec 10 19:42:10 crc kubenswrapper[4828]: I1210 19:42:10.855350 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jxhs" event={"ID":"ead62f92-39c7-46bb-be5d-f721322ee339","Type":"ContainerDied","Data":"b5081e46e877e4b88a775bf140bb153b368c60cbb91bf6e9c14ef3be4643017e"} Dec 10 19:42:10 crc kubenswrapper[4828]: I1210 19:42:10.855664 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jxhs" event={"ID":"ead62f92-39c7-46bb-be5d-f721322ee339","Type":"ContainerStarted","Data":"21c0f723a88cc3b0ca87807477d80d8bc59c311630a3634942e1310bb0f7fb99"} Dec 10 19:42:19 crc kubenswrapper[4828]: I1210 19:42:19.956193 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jxhs" event={"ID":"ead62f92-39c7-46bb-be5d-f721322ee339","Type":"ContainerStarted","Data":"0034d8bace6d6dcf98b4728113f838a24033d0509913c5055af688ca66f0ad8d"} Dec 10 19:42:21 crc kubenswrapper[4828]: I1210 19:42:21.979196 4828 generic.go:334] "Generic (PLEG): container finished" podID="ead62f92-39c7-46bb-be5d-f721322ee339" containerID="0034d8bace6d6dcf98b4728113f838a24033d0509913c5055af688ca66f0ad8d" exitCode=0 Dec 10 19:42:21 crc kubenswrapper[4828]: I1210 19:42:21.979271 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jxhs" event={"ID":"ead62f92-39c7-46bb-be5d-f721322ee339","Type":"ContainerDied","Data":"0034d8bace6d6dcf98b4728113f838a24033d0509913c5055af688ca66f0ad8d"} Dec 10 19:42:22 crc kubenswrapper[4828]: I1210 19:42:22.992413 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9jxhs" event={"ID":"ead62f92-39c7-46bb-be5d-f721322ee339","Type":"ContainerStarted","Data":"4963fee79a41a6cc83550a109c7b70530d308c35a045601c53f771dca7f0988e"} Dec 10 19:42:23 crc kubenswrapper[4828]: I1210 19:42:23.015848 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9jxhs" podStartSLOduration=2.280323607 podStartE2EDuration="14.015826403s" podCreationTimestamp="2025-12-10 19:42:09 +0000 UTC" firstStartedPulling="2025-12-10 19:42:10.857398094 +0000 UTC m=+2811.368009089" lastFinishedPulling="2025-12-10 19:42:22.59290088 +0000 UTC m=+2823.103511885" observedRunningTime="2025-12-10 19:42:23.007188553 +0000 UTC m=+2823.517799558" watchObservedRunningTime="2025-12-10 19:42:23.015826403 +0000 UTC m=+2823.526437408" Dec 10 19:42:30 crc kubenswrapper[4828]: I1210 19:42:30.034428 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9jxhs" Dec 10 19:42:30 crc kubenswrapper[4828]: I1210 19:42:30.035699 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9jxhs" Dec 10 19:42:30 crc kubenswrapper[4828]: I1210 19:42:30.113214 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9jxhs" Dec 10 19:42:30 crc kubenswrapper[4828]: I1210 19:42:30.164173 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9jxhs" Dec 10 19:42:30 crc kubenswrapper[4828]: I1210 19:42:30.237205 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9jxhs"] Dec 10 19:42:30 crc kubenswrapper[4828]: I1210 19:42:30.361875 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g7mhn"] Dec 10 19:42:30 crc kubenswrapper[4828]: I1210 19:42:30.362321 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g7mhn" podUID="3a8395da-9d4b-4d62-be88-dd60993239ef" containerName="registry-server" containerID="cri-o://350cfa1360e85b1bf666bba957451d1c0307b3ff6fde23a632bea378ad1eb706" gracePeriod=2 Dec 10 19:42:30 crc kubenswrapper[4828]: I1210 19:42:30.926884 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g7mhn" Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.006047 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4vqt\" (UniqueName: \"kubernetes.io/projected/3a8395da-9d4b-4d62-be88-dd60993239ef-kube-api-access-l4vqt\") pod \"3a8395da-9d4b-4d62-be88-dd60993239ef\" (UID: \"3a8395da-9d4b-4d62-be88-dd60993239ef\") " Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.006092 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a8395da-9d4b-4d62-be88-dd60993239ef-catalog-content\") pod \"3a8395da-9d4b-4d62-be88-dd60993239ef\" (UID: \"3a8395da-9d4b-4d62-be88-dd60993239ef\") " Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.006130 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a8395da-9d4b-4d62-be88-dd60993239ef-utilities\") pod \"3a8395da-9d4b-4d62-be88-dd60993239ef\" (UID: \"3a8395da-9d4b-4d62-be88-dd60993239ef\") " Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.007560 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a8395da-9d4b-4d62-be88-dd60993239ef-utilities" (OuterVolumeSpecName: "utilities") pod "3a8395da-9d4b-4d62-be88-dd60993239ef" (UID: "3a8395da-9d4b-4d62-be88-dd60993239ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.044017 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a8395da-9d4b-4d62-be88-dd60993239ef-kube-api-access-l4vqt" (OuterVolumeSpecName: "kube-api-access-l4vqt") pod "3a8395da-9d4b-4d62-be88-dd60993239ef" (UID: "3a8395da-9d4b-4d62-be88-dd60993239ef"). InnerVolumeSpecName "kube-api-access-l4vqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.087285 4828 generic.go:334] "Generic (PLEG): container finished" podID="3a8395da-9d4b-4d62-be88-dd60993239ef" containerID="350cfa1360e85b1bf666bba957451d1c0307b3ff6fde23a632bea378ad1eb706" exitCode=0 Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.088526 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g7mhn" Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.088700 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7mhn" event={"ID":"3a8395da-9d4b-4d62-be88-dd60993239ef","Type":"ContainerDied","Data":"350cfa1360e85b1bf666bba957451d1c0307b3ff6fde23a632bea378ad1eb706"} Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.089260 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g7mhn" event={"ID":"3a8395da-9d4b-4d62-be88-dd60993239ef","Type":"ContainerDied","Data":"c209090aacdafb46c3c3fa0b623955b9559ae1e94efdaf3cd696b4b659fcd32e"} Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.089349 4828 scope.go:117] "RemoveContainer" containerID="350cfa1360e85b1bf666bba957451d1c0307b3ff6fde23a632bea378ad1eb706" Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.109626 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4vqt\" (UniqueName: \"kubernetes.io/projected/3a8395da-9d4b-4d62-be88-dd60993239ef-kube-api-access-l4vqt\") on node \"crc\" DevicePath \"\"" Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.109838 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a8395da-9d4b-4d62-be88-dd60993239ef-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.121003 4828 scope.go:117] "RemoveContainer" containerID="8114cdcac4c1f2e74056f21cdf7c87a3ac9c5a9674e5a9d2598a81073a6010f8" Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.149063 4828 scope.go:117] "RemoveContainer" containerID="43c98e13d168a86eda789b6db1e1fb8f7c8967f151f0c5aa30db6dfbeb9cba45" Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.163921 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a8395da-9d4b-4d62-be88-dd60993239ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a8395da-9d4b-4d62-be88-dd60993239ef" (UID: "3a8395da-9d4b-4d62-be88-dd60993239ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.203858 4828 scope.go:117] "RemoveContainer" containerID="350cfa1360e85b1bf666bba957451d1c0307b3ff6fde23a632bea378ad1eb706" Dec 10 19:42:31 crc kubenswrapper[4828]: E1210 19:42:31.204239 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"350cfa1360e85b1bf666bba957451d1c0307b3ff6fde23a632bea378ad1eb706\": container with ID starting with 350cfa1360e85b1bf666bba957451d1c0307b3ff6fde23a632bea378ad1eb706 not found: ID does not exist" containerID="350cfa1360e85b1bf666bba957451d1c0307b3ff6fde23a632bea378ad1eb706" Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.204292 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"350cfa1360e85b1bf666bba957451d1c0307b3ff6fde23a632bea378ad1eb706"} err="failed to get container status \"350cfa1360e85b1bf666bba957451d1c0307b3ff6fde23a632bea378ad1eb706\": rpc error: code = NotFound desc = could not find container \"350cfa1360e85b1bf666bba957451d1c0307b3ff6fde23a632bea378ad1eb706\": container with ID starting with 350cfa1360e85b1bf666bba957451d1c0307b3ff6fde23a632bea378ad1eb706 not found: ID does not exist" Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.204331 4828 scope.go:117] "RemoveContainer" containerID="8114cdcac4c1f2e74056f21cdf7c87a3ac9c5a9674e5a9d2598a81073a6010f8" Dec 10 19:42:31 crc kubenswrapper[4828]: E1210 19:42:31.204664 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8114cdcac4c1f2e74056f21cdf7c87a3ac9c5a9674e5a9d2598a81073a6010f8\": container with ID starting with 8114cdcac4c1f2e74056f21cdf7c87a3ac9c5a9674e5a9d2598a81073a6010f8 not found: ID does not exist" containerID="8114cdcac4c1f2e74056f21cdf7c87a3ac9c5a9674e5a9d2598a81073a6010f8" Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.204760 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8114cdcac4c1f2e74056f21cdf7c87a3ac9c5a9674e5a9d2598a81073a6010f8"} err="failed to get container status \"8114cdcac4c1f2e74056f21cdf7c87a3ac9c5a9674e5a9d2598a81073a6010f8\": rpc error: code = NotFound desc = could not find container \"8114cdcac4c1f2e74056f21cdf7c87a3ac9c5a9674e5a9d2598a81073a6010f8\": container with ID starting with 8114cdcac4c1f2e74056f21cdf7c87a3ac9c5a9674e5a9d2598a81073a6010f8 not found: ID does not exist" Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.204945 4828 scope.go:117] "RemoveContainer" containerID="43c98e13d168a86eda789b6db1e1fb8f7c8967f151f0c5aa30db6dfbeb9cba45" Dec 10 19:42:31 crc kubenswrapper[4828]: E1210 19:42:31.205284 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43c98e13d168a86eda789b6db1e1fb8f7c8967f151f0c5aa30db6dfbeb9cba45\": container with ID starting with 43c98e13d168a86eda789b6db1e1fb8f7c8967f151f0c5aa30db6dfbeb9cba45 not found: ID does not exist" containerID="43c98e13d168a86eda789b6db1e1fb8f7c8967f151f0c5aa30db6dfbeb9cba45" Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.205322 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43c98e13d168a86eda789b6db1e1fb8f7c8967f151f0c5aa30db6dfbeb9cba45"} err="failed to get container status \"43c98e13d168a86eda789b6db1e1fb8f7c8967f151f0c5aa30db6dfbeb9cba45\": rpc error: code = NotFound desc = could not find container \"43c98e13d168a86eda789b6db1e1fb8f7c8967f151f0c5aa30db6dfbeb9cba45\": container with ID starting with 43c98e13d168a86eda789b6db1e1fb8f7c8967f151f0c5aa30db6dfbeb9cba45 not found: ID does not exist" Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.210852 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a8395da-9d4b-4d62-be88-dd60993239ef-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.445498 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g7mhn"] Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.468198 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g7mhn"] Dec 10 19:42:31 crc kubenswrapper[4828]: I1210 19:42:31.801988 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a8395da-9d4b-4d62-be88-dd60993239ef" path="/var/lib/kubelet/pods/3a8395da-9d4b-4d62-be88-dd60993239ef/volumes" Dec 10 19:42:51 crc kubenswrapper[4828]: I1210 19:42:51.230725 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:42:51 crc kubenswrapper[4828]: I1210 19:42:51.231334 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:43:21 crc kubenswrapper[4828]: I1210 19:43:21.230914 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:43:21 crc kubenswrapper[4828]: I1210 19:43:21.231560 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:43:45 crc kubenswrapper[4828]: I1210 19:43:45.143516 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9dl5l"] Dec 10 19:43:45 crc kubenswrapper[4828]: E1210 19:43:45.147125 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a8395da-9d4b-4d62-be88-dd60993239ef" containerName="extract-utilities" Dec 10 19:43:45 crc kubenswrapper[4828]: I1210 19:43:45.147145 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a8395da-9d4b-4d62-be88-dd60993239ef" containerName="extract-utilities" Dec 10 19:43:45 crc kubenswrapper[4828]: E1210 19:43:45.147177 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a8395da-9d4b-4d62-be88-dd60993239ef" containerName="extract-content" Dec 10 19:43:45 crc kubenswrapper[4828]: I1210 19:43:45.147184 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a8395da-9d4b-4d62-be88-dd60993239ef" containerName="extract-content" Dec 10 19:43:45 crc kubenswrapper[4828]: E1210 19:43:45.147201 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a8395da-9d4b-4d62-be88-dd60993239ef" containerName="registry-server" Dec 10 19:43:45 crc kubenswrapper[4828]: I1210 19:43:45.147207 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a8395da-9d4b-4d62-be88-dd60993239ef" containerName="registry-server" Dec 10 19:43:45 crc kubenswrapper[4828]: I1210 19:43:45.147439 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a8395da-9d4b-4d62-be88-dd60993239ef" containerName="registry-server" Dec 10 19:43:45 crc kubenswrapper[4828]: I1210 19:43:45.149161 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9dl5l" Dec 10 19:43:45 crc kubenswrapper[4828]: I1210 19:43:45.166894 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9dl5l"] Dec 10 19:43:45 crc kubenswrapper[4828]: I1210 19:43:45.270899 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80b0a158-165b-418a-b08d-0c81e303847c-utilities\") pod \"certified-operators-9dl5l\" (UID: \"80b0a158-165b-418a-b08d-0c81e303847c\") " pod="openshift-marketplace/certified-operators-9dl5l" Dec 10 19:43:45 crc kubenswrapper[4828]: I1210 19:43:45.270999 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52txs\" (UniqueName: \"kubernetes.io/projected/80b0a158-165b-418a-b08d-0c81e303847c-kube-api-access-52txs\") pod \"certified-operators-9dl5l\" (UID: \"80b0a158-165b-418a-b08d-0c81e303847c\") " pod="openshift-marketplace/certified-operators-9dl5l" Dec 10 19:43:45 crc kubenswrapper[4828]: I1210 19:43:45.271038 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80b0a158-165b-418a-b08d-0c81e303847c-catalog-content\") pod \"certified-operators-9dl5l\" (UID: \"80b0a158-165b-418a-b08d-0c81e303847c\") " pod="openshift-marketplace/certified-operators-9dl5l" Dec 10 19:43:45 crc kubenswrapper[4828]: I1210 19:43:45.375503 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80b0a158-165b-418a-b08d-0c81e303847c-utilities\") pod \"certified-operators-9dl5l\" (UID: \"80b0a158-165b-418a-b08d-0c81e303847c\") " pod="openshift-marketplace/certified-operators-9dl5l" Dec 10 19:43:45 crc kubenswrapper[4828]: I1210 19:43:45.375623 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52txs\" (UniqueName: \"kubernetes.io/projected/80b0a158-165b-418a-b08d-0c81e303847c-kube-api-access-52txs\") pod \"certified-operators-9dl5l\" (UID: \"80b0a158-165b-418a-b08d-0c81e303847c\") " pod="openshift-marketplace/certified-operators-9dl5l" Dec 10 19:43:45 crc kubenswrapper[4828]: I1210 19:43:45.375672 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80b0a158-165b-418a-b08d-0c81e303847c-catalog-content\") pod \"certified-operators-9dl5l\" (UID: \"80b0a158-165b-418a-b08d-0c81e303847c\") " pod="openshift-marketplace/certified-operators-9dl5l" Dec 10 19:43:45 crc kubenswrapper[4828]: I1210 19:43:45.376060 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80b0a158-165b-418a-b08d-0c81e303847c-utilities\") pod \"certified-operators-9dl5l\" (UID: \"80b0a158-165b-418a-b08d-0c81e303847c\") " pod="openshift-marketplace/certified-operators-9dl5l" Dec 10 19:43:45 crc kubenswrapper[4828]: I1210 19:43:45.376232 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80b0a158-165b-418a-b08d-0c81e303847c-catalog-content\") pod \"certified-operators-9dl5l\" (UID: \"80b0a158-165b-418a-b08d-0c81e303847c\") " pod="openshift-marketplace/certified-operators-9dl5l" Dec 10 19:43:45 crc kubenswrapper[4828]: I1210 19:43:45.412923 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52txs\" (UniqueName: \"kubernetes.io/projected/80b0a158-165b-418a-b08d-0c81e303847c-kube-api-access-52txs\") pod \"certified-operators-9dl5l\" (UID: \"80b0a158-165b-418a-b08d-0c81e303847c\") " pod="openshift-marketplace/certified-operators-9dl5l" Dec 10 19:43:45 crc kubenswrapper[4828]: I1210 19:43:45.498882 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9dl5l" Dec 10 19:43:46 crc kubenswrapper[4828]: I1210 19:43:46.069663 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9dl5l"] Dec 10 19:43:46 crc kubenswrapper[4828]: I1210 19:43:46.923483 4828 generic.go:334] "Generic (PLEG): container finished" podID="80b0a158-165b-418a-b08d-0c81e303847c" containerID="243190a8614ece3a1570406796373e89296cdd941aa051a93763687e810be096" exitCode=0 Dec 10 19:43:46 crc kubenswrapper[4828]: I1210 19:43:46.923857 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dl5l" event={"ID":"80b0a158-165b-418a-b08d-0c81e303847c","Type":"ContainerDied","Data":"243190a8614ece3a1570406796373e89296cdd941aa051a93763687e810be096"} Dec 10 19:43:46 crc kubenswrapper[4828]: I1210 19:43:46.923891 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dl5l" event={"ID":"80b0a158-165b-418a-b08d-0c81e303847c","Type":"ContainerStarted","Data":"3978ad49fc7eb918eafc195564fc07ad61baf54064f09829a7d959f6a2e7ffa2"} Dec 10 19:43:48 crc kubenswrapper[4828]: I1210 19:43:48.945284 4828 generic.go:334] "Generic (PLEG): container finished" podID="80b0a158-165b-418a-b08d-0c81e303847c" containerID="3fb8f6d9a5fdda8a8b104155386b39ebe854c3b3dd816f39b4fa004e82bcf6cb" exitCode=0 Dec 10 19:43:48 crc kubenswrapper[4828]: I1210 19:43:48.945373 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dl5l" event={"ID":"80b0a158-165b-418a-b08d-0c81e303847c","Type":"ContainerDied","Data":"3fb8f6d9a5fdda8a8b104155386b39ebe854c3b3dd816f39b4fa004e82bcf6cb"} Dec 10 19:43:49 crc kubenswrapper[4828]: I1210 19:43:49.965050 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dl5l" event={"ID":"80b0a158-165b-418a-b08d-0c81e303847c","Type":"ContainerStarted","Data":"381260f25350e061903de428cb01f80a435661e0ed54ec88051dfd6760faaa83"} Dec 10 19:43:50 crc kubenswrapper[4828]: I1210 19:43:50.006850 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9dl5l" podStartSLOduration=2.394899307 podStartE2EDuration="5.006828594s" podCreationTimestamp="2025-12-10 19:43:45 +0000 UTC" firstStartedPulling="2025-12-10 19:43:46.926762353 +0000 UTC m=+2907.437373358" lastFinishedPulling="2025-12-10 19:43:49.53869161 +0000 UTC m=+2910.049302645" observedRunningTime="2025-12-10 19:43:49.989217238 +0000 UTC m=+2910.499828293" watchObservedRunningTime="2025-12-10 19:43:50.006828594 +0000 UTC m=+2910.517439619" Dec 10 19:43:51 crc kubenswrapper[4828]: I1210 19:43:51.230573 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:43:51 crc kubenswrapper[4828]: I1210 19:43:51.230933 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:43:51 crc kubenswrapper[4828]: I1210 19:43:51.230987 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 19:43:51 crc kubenswrapper[4828]: I1210 19:43:51.231901 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:43:51 crc kubenswrapper[4828]: I1210 19:43:51.231957 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" gracePeriod=600 Dec 10 19:43:51 crc kubenswrapper[4828]: E1210 19:43:51.858920 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:43:51 crc kubenswrapper[4828]: I1210 19:43:51.990195 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" exitCode=0 Dec 10 19:43:51 crc kubenswrapper[4828]: I1210 19:43:51.990250 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83"} Dec 10 19:43:51 crc kubenswrapper[4828]: I1210 19:43:51.990322 4828 scope.go:117] "RemoveContainer" containerID="823c86952f20d947f08d0d8bb359ea62a6adbee8aa27350f3f768ba427671ebc" Dec 10 19:43:51 crc kubenswrapper[4828]: I1210 19:43:51.991243 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:43:51 crc kubenswrapper[4828]: E1210 19:43:51.991672 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:43:55 crc kubenswrapper[4828]: I1210 19:43:55.499279 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9dl5l" Dec 10 19:43:55 crc kubenswrapper[4828]: I1210 19:43:55.499843 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9dl5l" Dec 10 19:43:55 crc kubenswrapper[4828]: I1210 19:43:55.550281 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9dl5l" Dec 10 19:43:56 crc kubenswrapper[4828]: I1210 19:43:56.094860 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9dl5l" Dec 10 19:43:56 crc kubenswrapper[4828]: I1210 19:43:56.162365 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9dl5l"] Dec 10 19:43:58 crc kubenswrapper[4828]: I1210 19:43:58.067319 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9dl5l" podUID="80b0a158-165b-418a-b08d-0c81e303847c" containerName="registry-server" containerID="cri-o://381260f25350e061903de428cb01f80a435661e0ed54ec88051dfd6760faaa83" gracePeriod=2 Dec 10 19:43:58 crc kubenswrapper[4828]: I1210 19:43:58.549747 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9dl5l" Dec 10 19:43:58 crc kubenswrapper[4828]: I1210 19:43:58.609000 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80b0a158-165b-418a-b08d-0c81e303847c-catalog-content\") pod \"80b0a158-165b-418a-b08d-0c81e303847c\" (UID: \"80b0a158-165b-418a-b08d-0c81e303847c\") " Dec 10 19:43:58 crc kubenswrapper[4828]: I1210 19:43:58.609340 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52txs\" (UniqueName: \"kubernetes.io/projected/80b0a158-165b-418a-b08d-0c81e303847c-kube-api-access-52txs\") pod \"80b0a158-165b-418a-b08d-0c81e303847c\" (UID: \"80b0a158-165b-418a-b08d-0c81e303847c\") " Dec 10 19:43:58 crc kubenswrapper[4828]: I1210 19:43:58.609471 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80b0a158-165b-418a-b08d-0c81e303847c-utilities\") pod \"80b0a158-165b-418a-b08d-0c81e303847c\" (UID: \"80b0a158-165b-418a-b08d-0c81e303847c\") " Dec 10 19:43:58 crc kubenswrapper[4828]: I1210 19:43:58.611163 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80b0a158-165b-418a-b08d-0c81e303847c-utilities" (OuterVolumeSpecName: "utilities") pod "80b0a158-165b-418a-b08d-0c81e303847c" (UID: "80b0a158-165b-418a-b08d-0c81e303847c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:43:58 crc kubenswrapper[4828]: I1210 19:43:58.616657 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80b0a158-165b-418a-b08d-0c81e303847c-kube-api-access-52txs" (OuterVolumeSpecName: "kube-api-access-52txs") pod "80b0a158-165b-418a-b08d-0c81e303847c" (UID: "80b0a158-165b-418a-b08d-0c81e303847c"). InnerVolumeSpecName "kube-api-access-52txs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:43:58 crc kubenswrapper[4828]: I1210 19:43:58.662202 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80b0a158-165b-418a-b08d-0c81e303847c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80b0a158-165b-418a-b08d-0c81e303847c" (UID: "80b0a158-165b-418a-b08d-0c81e303847c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:43:58 crc kubenswrapper[4828]: I1210 19:43:58.714521 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80b0a158-165b-418a-b08d-0c81e303847c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:43:58 crc kubenswrapper[4828]: I1210 19:43:58.714566 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52txs\" (UniqueName: \"kubernetes.io/projected/80b0a158-165b-418a-b08d-0c81e303847c-kube-api-access-52txs\") on node \"crc\" DevicePath \"\"" Dec 10 19:43:58 crc kubenswrapper[4828]: I1210 19:43:58.714581 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80b0a158-165b-418a-b08d-0c81e303847c-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:43:59 crc kubenswrapper[4828]: I1210 19:43:59.078523 4828 generic.go:334] "Generic (PLEG): container finished" podID="80b0a158-165b-418a-b08d-0c81e303847c" containerID="381260f25350e061903de428cb01f80a435661e0ed54ec88051dfd6760faaa83" exitCode=0 Dec 10 19:43:59 crc kubenswrapper[4828]: I1210 19:43:59.078577 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dl5l" event={"ID":"80b0a158-165b-418a-b08d-0c81e303847c","Type":"ContainerDied","Data":"381260f25350e061903de428cb01f80a435661e0ed54ec88051dfd6760faaa83"} Dec 10 19:43:59 crc kubenswrapper[4828]: I1210 19:43:59.078588 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9dl5l" Dec 10 19:43:59 crc kubenswrapper[4828]: I1210 19:43:59.078614 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dl5l" event={"ID":"80b0a158-165b-418a-b08d-0c81e303847c","Type":"ContainerDied","Data":"3978ad49fc7eb918eafc195564fc07ad61baf54064f09829a7d959f6a2e7ffa2"} Dec 10 19:43:59 crc kubenswrapper[4828]: I1210 19:43:59.078639 4828 scope.go:117] "RemoveContainer" containerID="381260f25350e061903de428cb01f80a435661e0ed54ec88051dfd6760faaa83" Dec 10 19:43:59 crc kubenswrapper[4828]: I1210 19:43:59.099079 4828 scope.go:117] "RemoveContainer" containerID="3fb8f6d9a5fdda8a8b104155386b39ebe854c3b3dd816f39b4fa004e82bcf6cb" Dec 10 19:43:59 crc kubenswrapper[4828]: I1210 19:43:59.122549 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9dl5l"] Dec 10 19:43:59 crc kubenswrapper[4828]: I1210 19:43:59.133439 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9dl5l"] Dec 10 19:43:59 crc kubenswrapper[4828]: I1210 19:43:59.141147 4828 scope.go:117] "RemoveContainer" containerID="243190a8614ece3a1570406796373e89296cdd941aa051a93763687e810be096" Dec 10 19:43:59 crc kubenswrapper[4828]: I1210 19:43:59.182888 4828 scope.go:117] "RemoveContainer" containerID="381260f25350e061903de428cb01f80a435661e0ed54ec88051dfd6760faaa83" Dec 10 19:43:59 crc kubenswrapper[4828]: E1210 19:43:59.183560 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"381260f25350e061903de428cb01f80a435661e0ed54ec88051dfd6760faaa83\": container with ID starting with 381260f25350e061903de428cb01f80a435661e0ed54ec88051dfd6760faaa83 not found: ID does not exist" containerID="381260f25350e061903de428cb01f80a435661e0ed54ec88051dfd6760faaa83" Dec 10 19:43:59 crc kubenswrapper[4828]: I1210 19:43:59.183594 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"381260f25350e061903de428cb01f80a435661e0ed54ec88051dfd6760faaa83"} err="failed to get container status \"381260f25350e061903de428cb01f80a435661e0ed54ec88051dfd6760faaa83\": rpc error: code = NotFound desc = could not find container \"381260f25350e061903de428cb01f80a435661e0ed54ec88051dfd6760faaa83\": container with ID starting with 381260f25350e061903de428cb01f80a435661e0ed54ec88051dfd6760faaa83 not found: ID does not exist" Dec 10 19:43:59 crc kubenswrapper[4828]: I1210 19:43:59.183615 4828 scope.go:117] "RemoveContainer" containerID="3fb8f6d9a5fdda8a8b104155386b39ebe854c3b3dd816f39b4fa004e82bcf6cb" Dec 10 19:43:59 crc kubenswrapper[4828]: E1210 19:43:59.184024 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fb8f6d9a5fdda8a8b104155386b39ebe854c3b3dd816f39b4fa004e82bcf6cb\": container with ID starting with 3fb8f6d9a5fdda8a8b104155386b39ebe854c3b3dd816f39b4fa004e82bcf6cb not found: ID does not exist" containerID="3fb8f6d9a5fdda8a8b104155386b39ebe854c3b3dd816f39b4fa004e82bcf6cb" Dec 10 19:43:59 crc kubenswrapper[4828]: I1210 19:43:59.184048 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fb8f6d9a5fdda8a8b104155386b39ebe854c3b3dd816f39b4fa004e82bcf6cb"} err="failed to get container status \"3fb8f6d9a5fdda8a8b104155386b39ebe854c3b3dd816f39b4fa004e82bcf6cb\": rpc error: code = NotFound desc = could not find container \"3fb8f6d9a5fdda8a8b104155386b39ebe854c3b3dd816f39b4fa004e82bcf6cb\": container with ID starting with 3fb8f6d9a5fdda8a8b104155386b39ebe854c3b3dd816f39b4fa004e82bcf6cb not found: ID does not exist" Dec 10 19:43:59 crc kubenswrapper[4828]: I1210 19:43:59.184061 4828 scope.go:117] "RemoveContainer" containerID="243190a8614ece3a1570406796373e89296cdd941aa051a93763687e810be096" Dec 10 19:43:59 crc kubenswrapper[4828]: E1210 19:43:59.184320 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"243190a8614ece3a1570406796373e89296cdd941aa051a93763687e810be096\": container with ID starting with 243190a8614ece3a1570406796373e89296cdd941aa051a93763687e810be096 not found: ID does not exist" containerID="243190a8614ece3a1570406796373e89296cdd941aa051a93763687e810be096" Dec 10 19:43:59 crc kubenswrapper[4828]: I1210 19:43:59.184344 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"243190a8614ece3a1570406796373e89296cdd941aa051a93763687e810be096"} err="failed to get container status \"243190a8614ece3a1570406796373e89296cdd941aa051a93763687e810be096\": rpc error: code = NotFound desc = could not find container \"243190a8614ece3a1570406796373e89296cdd941aa051a93763687e810be096\": container with ID starting with 243190a8614ece3a1570406796373e89296cdd941aa051a93763687e810be096 not found: ID does not exist" Dec 10 19:43:59 crc kubenswrapper[4828]: I1210 19:43:59.809404 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80b0a158-165b-418a-b08d-0c81e303847c" path="/var/lib/kubelet/pods/80b0a158-165b-418a-b08d-0c81e303847c/volumes" Dec 10 19:44:05 crc kubenswrapper[4828]: I1210 19:44:05.789254 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:44:05 crc kubenswrapper[4828]: E1210 19:44:05.790107 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:44:16 crc kubenswrapper[4828]: I1210 19:44:16.789189 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:44:16 crc kubenswrapper[4828]: E1210 19:44:16.790317 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.048178 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4nflk"] Dec 10 19:44:27 crc kubenswrapper[4828]: E1210 19:44:27.049297 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b0a158-165b-418a-b08d-0c81e303847c" containerName="extract-content" Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.049314 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b0a158-165b-418a-b08d-0c81e303847c" containerName="extract-content" Dec 10 19:44:27 crc kubenswrapper[4828]: E1210 19:44:27.049332 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b0a158-165b-418a-b08d-0c81e303847c" containerName="registry-server" Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.049338 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b0a158-165b-418a-b08d-0c81e303847c" containerName="registry-server" Dec 10 19:44:27 crc kubenswrapper[4828]: E1210 19:44:27.049361 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80b0a158-165b-418a-b08d-0c81e303847c" containerName="extract-utilities" Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.049367 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="80b0a158-165b-418a-b08d-0c81e303847c" containerName="extract-utilities" Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.049861 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="80b0a158-165b-418a-b08d-0c81e303847c" containerName="registry-server" Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.061669 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4nflk" Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.073720 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4nflk"] Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.098865 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-utilities\") pod \"redhat-marketplace-4nflk\" (UID: \"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268\") " pod="openshift-marketplace/redhat-marketplace-4nflk" Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.099045 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-catalog-content\") pod \"redhat-marketplace-4nflk\" (UID: \"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268\") " pod="openshift-marketplace/redhat-marketplace-4nflk" Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.099095 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h289d\" (UniqueName: \"kubernetes.io/projected/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-kube-api-access-h289d\") pod \"redhat-marketplace-4nflk\" (UID: \"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268\") " pod="openshift-marketplace/redhat-marketplace-4nflk" Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.200770 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-catalog-content\") pod \"redhat-marketplace-4nflk\" (UID: \"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268\") " pod="openshift-marketplace/redhat-marketplace-4nflk" Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.201153 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h289d\" (UniqueName: \"kubernetes.io/projected/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-kube-api-access-h289d\") pod \"redhat-marketplace-4nflk\" (UID: \"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268\") " pod="openshift-marketplace/redhat-marketplace-4nflk" Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.201299 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-utilities\") pod \"redhat-marketplace-4nflk\" (UID: \"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268\") " pod="openshift-marketplace/redhat-marketplace-4nflk" Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.201303 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-catalog-content\") pod \"redhat-marketplace-4nflk\" (UID: \"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268\") " pod="openshift-marketplace/redhat-marketplace-4nflk" Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.201524 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-utilities\") pod \"redhat-marketplace-4nflk\" (UID: \"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268\") " pod="openshift-marketplace/redhat-marketplace-4nflk" Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.224554 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h289d\" (UniqueName: \"kubernetes.io/projected/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-kube-api-access-h289d\") pod \"redhat-marketplace-4nflk\" (UID: \"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268\") " pod="openshift-marketplace/redhat-marketplace-4nflk" Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.403748 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4nflk" Dec 10 19:44:27 crc kubenswrapper[4828]: I1210 19:44:27.889658 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4nflk"] Dec 10 19:44:28 crc kubenswrapper[4828]: I1210 19:44:28.374607 4828 generic.go:334] "Generic (PLEG): container finished" podID="ab49d0b6-c95f-4b20-8c2a-a9978d8c5268" containerID="c5293a72b74cb84452035dfe429b313e79a67193c071e395260c01fba16725da" exitCode=0 Dec 10 19:44:28 crc kubenswrapper[4828]: I1210 19:44:28.374688 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4nflk" event={"ID":"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268","Type":"ContainerDied","Data":"c5293a72b74cb84452035dfe429b313e79a67193c071e395260c01fba16725da"} Dec 10 19:44:28 crc kubenswrapper[4828]: I1210 19:44:28.375702 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4nflk" event={"ID":"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268","Type":"ContainerStarted","Data":"d641667b3cf9032c35d0c5999ef1ae8e36e3577411f69ee62b193ec626fd71e4"} Dec 10 19:44:30 crc kubenswrapper[4828]: I1210 19:44:30.403665 4828 generic.go:334] "Generic (PLEG): container finished" podID="ab49d0b6-c95f-4b20-8c2a-a9978d8c5268" containerID="ebc75f09b7ef0c5d225cfcdbf4b567ddb8ad949351480ef2c2b0733d49511466" exitCode=0 Dec 10 19:44:30 crc kubenswrapper[4828]: I1210 19:44:30.403774 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4nflk" event={"ID":"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268","Type":"ContainerDied","Data":"ebc75f09b7ef0c5d225cfcdbf4b567ddb8ad949351480ef2c2b0733d49511466"} Dec 10 19:44:30 crc kubenswrapper[4828]: I1210 19:44:30.789549 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:44:30 crc kubenswrapper[4828]: E1210 19:44:30.790201 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:44:32 crc kubenswrapper[4828]: I1210 19:44:32.427083 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4nflk" event={"ID":"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268","Type":"ContainerStarted","Data":"d6a0bd60663cc193dfed76de6ea51c20a7249d6efb1a0914bb2ed2aa37753029"} Dec 10 19:44:32 crc kubenswrapper[4828]: I1210 19:44:32.452792 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4nflk" podStartSLOduration=2.351883193 podStartE2EDuration="5.452660252s" podCreationTimestamp="2025-12-10 19:44:27 +0000 UTC" firstStartedPulling="2025-12-10 19:44:28.3761976 +0000 UTC m=+2948.886808605" lastFinishedPulling="2025-12-10 19:44:31.476974659 +0000 UTC m=+2951.987585664" observedRunningTime="2025-12-10 19:44:32.444723062 +0000 UTC m=+2952.955334067" watchObservedRunningTime="2025-12-10 19:44:32.452660252 +0000 UTC m=+2952.963271257" Dec 10 19:44:34 crc kubenswrapper[4828]: I1210 19:44:34.446688 4828 generic.go:334] "Generic (PLEG): container finished" podID="7dc2751a-2c10-409f-8313-f6e78fb87ba3" containerID="0352a738064fef72110c449470c67b89f3d647ef0fbcf33d751c501457d0b17c" exitCode=0 Dec 10 19:44:34 crc kubenswrapper[4828]: I1210 19:44:34.446778 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" event={"ID":"7dc2751a-2c10-409f-8313-f6e78fb87ba3","Type":"ContainerDied","Data":"0352a738064fef72110c449470c67b89f3d647ef0fbcf33d751c501457d0b17c"} Dec 10 19:44:35 crc kubenswrapper[4828]: I1210 19:44:35.948095 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.018749 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgwf5\" (UniqueName: \"kubernetes.io/projected/7dc2751a-2c10-409f-8313-f6e78fb87ba3-kube-api-access-pgwf5\") pod \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.018855 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-migration-ssh-key-0\") pod \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.018990 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-cell1-compute-config-0\") pod \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.019054 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-extra-config-0\") pod \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.019098 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-ssh-key\") pod \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.019137 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-migration-ssh-key-1\") pod \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.019210 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-combined-ca-bundle\") pod \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.019419 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-inventory\") pod \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.019453 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-cell1-compute-config-1\") pod \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\" (UID: \"7dc2751a-2c10-409f-8313-f6e78fb87ba3\") " Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.035136 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "7dc2751a-2c10-409f-8313-f6e78fb87ba3" (UID: "7dc2751a-2c10-409f-8313-f6e78fb87ba3"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.037026 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dc2751a-2c10-409f-8313-f6e78fb87ba3-kube-api-access-pgwf5" (OuterVolumeSpecName: "kube-api-access-pgwf5") pod "7dc2751a-2c10-409f-8313-f6e78fb87ba3" (UID: "7dc2751a-2c10-409f-8313-f6e78fb87ba3"). InnerVolumeSpecName "kube-api-access-pgwf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.057509 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "7dc2751a-2c10-409f-8313-f6e78fb87ba3" (UID: "7dc2751a-2c10-409f-8313-f6e78fb87ba3"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.059030 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-inventory" (OuterVolumeSpecName: "inventory") pod "7dc2751a-2c10-409f-8313-f6e78fb87ba3" (UID: "7dc2751a-2c10-409f-8313-f6e78fb87ba3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.060207 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7dc2751a-2c10-409f-8313-f6e78fb87ba3" (UID: "7dc2751a-2c10-409f-8313-f6e78fb87ba3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.060895 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "7dc2751a-2c10-409f-8313-f6e78fb87ba3" (UID: "7dc2751a-2c10-409f-8313-f6e78fb87ba3"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.066149 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "7dc2751a-2c10-409f-8313-f6e78fb87ba3" (UID: "7dc2751a-2c10-409f-8313-f6e78fb87ba3"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.068718 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "7dc2751a-2c10-409f-8313-f6e78fb87ba3" (UID: "7dc2751a-2c10-409f-8313-f6e78fb87ba3"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.076199 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "7dc2751a-2c10-409f-8313-f6e78fb87ba3" (UID: "7dc2751a-2c10-409f-8313-f6e78fb87ba3"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.123013 4828 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.123058 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.123069 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgwf5\" (UniqueName: \"kubernetes.io/projected/7dc2751a-2c10-409f-8313-f6e78fb87ba3-kube-api-access-pgwf5\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.123080 4828 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.123089 4828 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.123099 4828 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.123108 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.123116 4828 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.123125 4828 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dc2751a-2c10-409f-8313-f6e78fb87ba3-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.470906 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" event={"ID":"7dc2751a-2c10-409f-8313-f6e78fb87ba3","Type":"ContainerDied","Data":"d2d9ac3b54116080316cf2edaf5f90061b13675385d5f55968504e7aa65184ee"} Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.470967 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2d9ac3b54116080316cf2edaf5f90061b13675385d5f55968504e7aa65184ee" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.471029 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-475xn" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.553262 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs"] Dec 10 19:44:36 crc kubenswrapper[4828]: E1210 19:44:36.553745 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dc2751a-2c10-409f-8313-f6e78fb87ba3" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.553765 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dc2751a-2c10-409f-8313-f6e78fb87ba3" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.554027 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dc2751a-2c10-409f-8313-f6e78fb87ba3" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.554818 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.557419 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.557540 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.557419 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.558016 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.558566 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.567100 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs"] Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.634003 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.634072 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.634189 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.634306 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgbt2\" (UniqueName: \"kubernetes.io/projected/f5034a0e-652b-414a-a36f-e7f4f34da750-kube-api-access-kgbt2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.634393 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.634557 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.634722 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.736945 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.737055 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.737111 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.737136 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.737173 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.737214 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgbt2\" (UniqueName: \"kubernetes.io/projected/f5034a0e-652b-414a-a36f-e7f4f34da750-kube-api-access-kgbt2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.737263 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.743203 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.743259 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.743669 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.744566 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.752083 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.758580 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.760713 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgbt2\" (UniqueName: \"kubernetes.io/projected/f5034a0e-652b-414a-a36f-e7f4f34da750-kube-api-access-kgbt2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-csrqs\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:36 crc kubenswrapper[4828]: I1210 19:44:36.875329 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:44:37 crc kubenswrapper[4828]: I1210 19:44:37.405434 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4nflk" Dec 10 19:44:37 crc kubenswrapper[4828]: I1210 19:44:37.406054 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4nflk" Dec 10 19:44:37 crc kubenswrapper[4828]: I1210 19:44:37.406877 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs"] Dec 10 19:44:37 crc kubenswrapper[4828]: I1210 19:44:37.464812 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4nflk" Dec 10 19:44:37 crc kubenswrapper[4828]: I1210 19:44:37.487280 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" event={"ID":"f5034a0e-652b-414a-a36f-e7f4f34da750","Type":"ContainerStarted","Data":"2f833b94c3ffdfb3aeb1272c4ecfd9a3cfcd9d7e463ac33fe059eed3c54b81cd"} Dec 10 19:44:37 crc kubenswrapper[4828]: I1210 19:44:37.533294 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4nflk" Dec 10 19:44:38 crc kubenswrapper[4828]: I1210 19:44:38.499365 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" event={"ID":"f5034a0e-652b-414a-a36f-e7f4f34da750","Type":"ContainerStarted","Data":"346abaccfc8d2cdbd3379f61f60e5bedca51d1707b7c0d22103d357d4034d306"} Dec 10 19:44:39 crc kubenswrapper[4828]: I1210 19:44:39.016680 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" podStartSLOduration=2.560998792 podStartE2EDuration="3.016660477s" podCreationTimestamp="2025-12-10 19:44:36 +0000 UTC" firstStartedPulling="2025-12-10 19:44:37.412315678 +0000 UTC m=+2957.922926683" lastFinishedPulling="2025-12-10 19:44:37.867977363 +0000 UTC m=+2958.378588368" observedRunningTime="2025-12-10 19:44:38.530243288 +0000 UTC m=+2959.040854323" watchObservedRunningTime="2025-12-10 19:44:39.016660477 +0000 UTC m=+2959.527271482" Dec 10 19:44:39 crc kubenswrapper[4828]: I1210 19:44:39.031912 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4nflk"] Dec 10 19:44:39 crc kubenswrapper[4828]: I1210 19:44:39.512041 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4nflk" podUID="ab49d0b6-c95f-4b20-8c2a-a9978d8c5268" containerName="registry-server" containerID="cri-o://d6a0bd60663cc193dfed76de6ea51c20a7249d6efb1a0914bb2ed2aa37753029" gracePeriod=2 Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.050355 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4nflk" Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.129828 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-catalog-content\") pod \"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268\" (UID: \"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268\") " Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.130340 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-utilities\") pod \"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268\" (UID: \"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268\") " Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.130549 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h289d\" (UniqueName: \"kubernetes.io/projected/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-kube-api-access-h289d\") pod \"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268\" (UID: \"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268\") " Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.131222 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-utilities" (OuterVolumeSpecName: "utilities") pod "ab49d0b6-c95f-4b20-8c2a-a9978d8c5268" (UID: "ab49d0b6-c95f-4b20-8c2a-a9978d8c5268"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.136745 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-kube-api-access-h289d" (OuterVolumeSpecName: "kube-api-access-h289d") pod "ab49d0b6-c95f-4b20-8c2a-a9978d8c5268" (UID: "ab49d0b6-c95f-4b20-8c2a-a9978d8c5268"). InnerVolumeSpecName "kube-api-access-h289d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.152310 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ab49d0b6-c95f-4b20-8c2a-a9978d8c5268" (UID: "ab49d0b6-c95f-4b20-8c2a-a9978d8c5268"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.233747 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.233791 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h289d\" (UniqueName: \"kubernetes.io/projected/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-kube-api-access-h289d\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.233817 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.523525 4828 generic.go:334] "Generic (PLEG): container finished" podID="ab49d0b6-c95f-4b20-8c2a-a9978d8c5268" containerID="d6a0bd60663cc193dfed76de6ea51c20a7249d6efb1a0914bb2ed2aa37753029" exitCode=0 Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.523570 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4nflk" event={"ID":"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268","Type":"ContainerDied","Data":"d6a0bd60663cc193dfed76de6ea51c20a7249d6efb1a0914bb2ed2aa37753029"} Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.523593 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4nflk" Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.523614 4828 scope.go:117] "RemoveContainer" containerID="d6a0bd60663cc193dfed76de6ea51c20a7249d6efb1a0914bb2ed2aa37753029" Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.523600 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4nflk" event={"ID":"ab49d0b6-c95f-4b20-8c2a-a9978d8c5268","Type":"ContainerDied","Data":"d641667b3cf9032c35d0c5999ef1ae8e36e3577411f69ee62b193ec626fd71e4"} Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.558066 4828 scope.go:117] "RemoveContainer" containerID="ebc75f09b7ef0c5d225cfcdbf4b567ddb8ad949351480ef2c2b0733d49511466" Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.563112 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4nflk"] Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.576982 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4nflk"] Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.582506 4828 scope.go:117] "RemoveContainer" containerID="c5293a72b74cb84452035dfe429b313e79a67193c071e395260c01fba16725da" Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.642754 4828 scope.go:117] "RemoveContainer" containerID="d6a0bd60663cc193dfed76de6ea51c20a7249d6efb1a0914bb2ed2aa37753029" Dec 10 19:44:40 crc kubenswrapper[4828]: E1210 19:44:40.643247 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6a0bd60663cc193dfed76de6ea51c20a7249d6efb1a0914bb2ed2aa37753029\": container with ID starting with d6a0bd60663cc193dfed76de6ea51c20a7249d6efb1a0914bb2ed2aa37753029 not found: ID does not exist" containerID="d6a0bd60663cc193dfed76de6ea51c20a7249d6efb1a0914bb2ed2aa37753029" Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.643281 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6a0bd60663cc193dfed76de6ea51c20a7249d6efb1a0914bb2ed2aa37753029"} err="failed to get container status \"d6a0bd60663cc193dfed76de6ea51c20a7249d6efb1a0914bb2ed2aa37753029\": rpc error: code = NotFound desc = could not find container \"d6a0bd60663cc193dfed76de6ea51c20a7249d6efb1a0914bb2ed2aa37753029\": container with ID starting with d6a0bd60663cc193dfed76de6ea51c20a7249d6efb1a0914bb2ed2aa37753029 not found: ID does not exist" Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.643304 4828 scope.go:117] "RemoveContainer" containerID="ebc75f09b7ef0c5d225cfcdbf4b567ddb8ad949351480ef2c2b0733d49511466" Dec 10 19:44:40 crc kubenswrapper[4828]: E1210 19:44:40.643638 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebc75f09b7ef0c5d225cfcdbf4b567ddb8ad949351480ef2c2b0733d49511466\": container with ID starting with ebc75f09b7ef0c5d225cfcdbf4b567ddb8ad949351480ef2c2b0733d49511466 not found: ID does not exist" containerID="ebc75f09b7ef0c5d225cfcdbf4b567ddb8ad949351480ef2c2b0733d49511466" Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.643663 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebc75f09b7ef0c5d225cfcdbf4b567ddb8ad949351480ef2c2b0733d49511466"} err="failed to get container status \"ebc75f09b7ef0c5d225cfcdbf4b567ddb8ad949351480ef2c2b0733d49511466\": rpc error: code = NotFound desc = could not find container \"ebc75f09b7ef0c5d225cfcdbf4b567ddb8ad949351480ef2c2b0733d49511466\": container with ID starting with ebc75f09b7ef0c5d225cfcdbf4b567ddb8ad949351480ef2c2b0733d49511466 not found: ID does not exist" Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.643678 4828 scope.go:117] "RemoveContainer" containerID="c5293a72b74cb84452035dfe429b313e79a67193c071e395260c01fba16725da" Dec 10 19:44:40 crc kubenswrapper[4828]: E1210 19:44:40.644232 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5293a72b74cb84452035dfe429b313e79a67193c071e395260c01fba16725da\": container with ID starting with c5293a72b74cb84452035dfe429b313e79a67193c071e395260c01fba16725da not found: ID does not exist" containerID="c5293a72b74cb84452035dfe429b313e79a67193c071e395260c01fba16725da" Dec 10 19:44:40 crc kubenswrapper[4828]: I1210 19:44:40.644254 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5293a72b74cb84452035dfe429b313e79a67193c071e395260c01fba16725da"} err="failed to get container status \"c5293a72b74cb84452035dfe429b313e79a67193c071e395260c01fba16725da\": rpc error: code = NotFound desc = could not find container \"c5293a72b74cb84452035dfe429b313e79a67193c071e395260c01fba16725da\": container with ID starting with c5293a72b74cb84452035dfe429b313e79a67193c071e395260c01fba16725da not found: ID does not exist" Dec 10 19:44:41 crc kubenswrapper[4828]: I1210 19:44:41.788821 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:44:41 crc kubenswrapper[4828]: E1210 19:44:41.789396 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:44:41 crc kubenswrapper[4828]: I1210 19:44:41.804561 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab49d0b6-c95f-4b20-8c2a-a9978d8c5268" path="/var/lib/kubelet/pods/ab49d0b6-c95f-4b20-8c2a-a9978d8c5268/volumes" Dec 10 19:44:52 crc kubenswrapper[4828]: I1210 19:44:52.788989 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:44:52 crc kubenswrapper[4828]: E1210 19:44:52.789810 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.157699 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr"] Dec 10 19:45:00 crc kubenswrapper[4828]: E1210 19:45:00.158867 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab49d0b6-c95f-4b20-8c2a-a9978d8c5268" containerName="extract-content" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.158886 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab49d0b6-c95f-4b20-8c2a-a9978d8c5268" containerName="extract-content" Dec 10 19:45:00 crc kubenswrapper[4828]: E1210 19:45:00.158946 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab49d0b6-c95f-4b20-8c2a-a9978d8c5268" containerName="extract-utilities" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.158962 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab49d0b6-c95f-4b20-8c2a-a9978d8c5268" containerName="extract-utilities" Dec 10 19:45:00 crc kubenswrapper[4828]: E1210 19:45:00.158998 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab49d0b6-c95f-4b20-8c2a-a9978d8c5268" containerName="registry-server" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.159006 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab49d0b6-c95f-4b20-8c2a-a9978d8c5268" containerName="registry-server" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.159301 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab49d0b6-c95f-4b20-8c2a-a9978d8c5268" containerName="registry-server" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.160350 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.162972 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.166039 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.185338 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr"] Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.284509 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f445c4c6-49ea-483f-b20c-22fb12e316ad-secret-volume\") pod \"collect-profiles-29423265-s7hpr\" (UID: \"f445c4c6-49ea-483f-b20c-22fb12e316ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.284577 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f445c4c6-49ea-483f-b20c-22fb12e316ad-config-volume\") pod \"collect-profiles-29423265-s7hpr\" (UID: \"f445c4c6-49ea-483f-b20c-22fb12e316ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.284606 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z8fv\" (UniqueName: \"kubernetes.io/projected/f445c4c6-49ea-483f-b20c-22fb12e316ad-kube-api-access-7z8fv\") pod \"collect-profiles-29423265-s7hpr\" (UID: \"f445c4c6-49ea-483f-b20c-22fb12e316ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.386293 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f445c4c6-49ea-483f-b20c-22fb12e316ad-secret-volume\") pod \"collect-profiles-29423265-s7hpr\" (UID: \"f445c4c6-49ea-483f-b20c-22fb12e316ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.386369 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f445c4c6-49ea-483f-b20c-22fb12e316ad-config-volume\") pod \"collect-profiles-29423265-s7hpr\" (UID: \"f445c4c6-49ea-483f-b20c-22fb12e316ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.386396 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z8fv\" (UniqueName: \"kubernetes.io/projected/f445c4c6-49ea-483f-b20c-22fb12e316ad-kube-api-access-7z8fv\") pod \"collect-profiles-29423265-s7hpr\" (UID: \"f445c4c6-49ea-483f-b20c-22fb12e316ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.387406 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f445c4c6-49ea-483f-b20c-22fb12e316ad-config-volume\") pod \"collect-profiles-29423265-s7hpr\" (UID: \"f445c4c6-49ea-483f-b20c-22fb12e316ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.395418 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f445c4c6-49ea-483f-b20c-22fb12e316ad-secret-volume\") pod \"collect-profiles-29423265-s7hpr\" (UID: \"f445c4c6-49ea-483f-b20c-22fb12e316ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.405482 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z8fv\" (UniqueName: \"kubernetes.io/projected/f445c4c6-49ea-483f-b20c-22fb12e316ad-kube-api-access-7z8fv\") pod \"collect-profiles-29423265-s7hpr\" (UID: \"f445c4c6-49ea-483f-b20c-22fb12e316ad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.483879 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr" Dec 10 19:45:00 crc kubenswrapper[4828]: I1210 19:45:00.925450 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr"] Dec 10 19:45:01 crc kubenswrapper[4828]: I1210 19:45:01.755727 4828 generic.go:334] "Generic (PLEG): container finished" podID="f445c4c6-49ea-483f-b20c-22fb12e316ad" containerID="1960712b3a4058cacb3039abfd7810ef09e544bd8f9ef480338e2450963e71e6" exitCode=0 Dec 10 19:45:01 crc kubenswrapper[4828]: I1210 19:45:01.755788 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr" event={"ID":"f445c4c6-49ea-483f-b20c-22fb12e316ad","Type":"ContainerDied","Data":"1960712b3a4058cacb3039abfd7810ef09e544bd8f9ef480338e2450963e71e6"} Dec 10 19:45:01 crc kubenswrapper[4828]: I1210 19:45:01.756039 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr" event={"ID":"f445c4c6-49ea-483f-b20c-22fb12e316ad","Type":"ContainerStarted","Data":"5ede9fdfd062dac6385b05be41e7426debca8cbbe2bcb4fb71f8695fdd09d1a5"} Dec 10 19:45:03 crc kubenswrapper[4828]: I1210 19:45:03.161782 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr" Dec 10 19:45:03 crc kubenswrapper[4828]: I1210 19:45:03.356228 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f445c4c6-49ea-483f-b20c-22fb12e316ad-secret-volume\") pod \"f445c4c6-49ea-483f-b20c-22fb12e316ad\" (UID: \"f445c4c6-49ea-483f-b20c-22fb12e316ad\") " Dec 10 19:45:03 crc kubenswrapper[4828]: I1210 19:45:03.356291 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7z8fv\" (UniqueName: \"kubernetes.io/projected/f445c4c6-49ea-483f-b20c-22fb12e316ad-kube-api-access-7z8fv\") pod \"f445c4c6-49ea-483f-b20c-22fb12e316ad\" (UID: \"f445c4c6-49ea-483f-b20c-22fb12e316ad\") " Dec 10 19:45:03 crc kubenswrapper[4828]: I1210 19:45:03.356339 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f445c4c6-49ea-483f-b20c-22fb12e316ad-config-volume\") pod \"f445c4c6-49ea-483f-b20c-22fb12e316ad\" (UID: \"f445c4c6-49ea-483f-b20c-22fb12e316ad\") " Dec 10 19:45:03 crc kubenswrapper[4828]: I1210 19:45:03.357118 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f445c4c6-49ea-483f-b20c-22fb12e316ad-config-volume" (OuterVolumeSpecName: "config-volume") pod "f445c4c6-49ea-483f-b20c-22fb12e316ad" (UID: "f445c4c6-49ea-483f-b20c-22fb12e316ad"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 19:45:03 crc kubenswrapper[4828]: I1210 19:45:03.364027 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f445c4c6-49ea-483f-b20c-22fb12e316ad-kube-api-access-7z8fv" (OuterVolumeSpecName: "kube-api-access-7z8fv") pod "f445c4c6-49ea-483f-b20c-22fb12e316ad" (UID: "f445c4c6-49ea-483f-b20c-22fb12e316ad"). InnerVolumeSpecName "kube-api-access-7z8fv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:45:03 crc kubenswrapper[4828]: I1210 19:45:03.363654 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f445c4c6-49ea-483f-b20c-22fb12e316ad-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f445c4c6-49ea-483f-b20c-22fb12e316ad" (UID: "f445c4c6-49ea-483f-b20c-22fb12e316ad"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:45:03 crc kubenswrapper[4828]: I1210 19:45:03.461416 4828 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f445c4c6-49ea-483f-b20c-22fb12e316ad-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 19:45:03 crc kubenswrapper[4828]: I1210 19:45:03.461457 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7z8fv\" (UniqueName: \"kubernetes.io/projected/f445c4c6-49ea-483f-b20c-22fb12e316ad-kube-api-access-7z8fv\") on node \"crc\" DevicePath \"\"" Dec 10 19:45:03 crc kubenswrapper[4828]: I1210 19:45:03.461471 4828 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f445c4c6-49ea-483f-b20c-22fb12e316ad-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 19:45:03 crc kubenswrapper[4828]: I1210 19:45:03.787243 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr" event={"ID":"f445c4c6-49ea-483f-b20c-22fb12e316ad","Type":"ContainerDied","Data":"5ede9fdfd062dac6385b05be41e7426debca8cbbe2bcb4fb71f8695fdd09d1a5"} Dec 10 19:45:03 crc kubenswrapper[4828]: I1210 19:45:03.788293 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ede9fdfd062dac6385b05be41e7426debca8cbbe2bcb4fb71f8695fdd09d1a5" Dec 10 19:45:03 crc kubenswrapper[4828]: I1210 19:45:03.787287 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr" Dec 10 19:45:03 crc kubenswrapper[4828]: I1210 19:45:03.806054 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:45:03 crc kubenswrapper[4828]: E1210 19:45:03.806441 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:45:04 crc kubenswrapper[4828]: I1210 19:45:04.237053 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd"] Dec 10 19:45:04 crc kubenswrapper[4828]: I1210 19:45:04.249023 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423220-8r8rd"] Dec 10 19:45:05 crc kubenswrapper[4828]: I1210 19:45:05.803055 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22097d8d-0d50-4948-984b-4ea71356df75" path="/var/lib/kubelet/pods/22097d8d-0d50-4948-984b-4ea71356df75/volumes" Dec 10 19:45:14 crc kubenswrapper[4828]: I1210 19:45:14.790451 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:45:14 crc kubenswrapper[4828]: E1210 19:45:14.791360 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:45:26 crc kubenswrapper[4828]: I1210 19:45:26.788882 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:45:26 crc kubenswrapper[4828]: E1210 19:45:26.789827 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:45:33 crc kubenswrapper[4828]: I1210 19:45:33.496131 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c8fzq"] Dec 10 19:45:33 crc kubenswrapper[4828]: E1210 19:45:33.497307 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f445c4c6-49ea-483f-b20c-22fb12e316ad" containerName="collect-profiles" Dec 10 19:45:33 crc kubenswrapper[4828]: I1210 19:45:33.497321 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="f445c4c6-49ea-483f-b20c-22fb12e316ad" containerName="collect-profiles" Dec 10 19:45:33 crc kubenswrapper[4828]: I1210 19:45:33.497529 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="f445c4c6-49ea-483f-b20c-22fb12e316ad" containerName="collect-profiles" Dec 10 19:45:33 crc kubenswrapper[4828]: I1210 19:45:33.499277 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c8fzq" Dec 10 19:45:33 crc kubenswrapper[4828]: I1210 19:45:33.514615 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c8fzq"] Dec 10 19:45:33 crc kubenswrapper[4828]: I1210 19:45:33.582922 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce15f26e-387c-48a0-86af-1cd73371bc7a-utilities\") pod \"community-operators-c8fzq\" (UID: \"ce15f26e-387c-48a0-86af-1cd73371bc7a\") " pod="openshift-marketplace/community-operators-c8fzq" Dec 10 19:45:33 crc kubenswrapper[4828]: I1210 19:45:33.582985 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtq9m\" (UniqueName: \"kubernetes.io/projected/ce15f26e-387c-48a0-86af-1cd73371bc7a-kube-api-access-rtq9m\") pod \"community-operators-c8fzq\" (UID: \"ce15f26e-387c-48a0-86af-1cd73371bc7a\") " pod="openshift-marketplace/community-operators-c8fzq" Dec 10 19:45:33 crc kubenswrapper[4828]: I1210 19:45:33.583064 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce15f26e-387c-48a0-86af-1cd73371bc7a-catalog-content\") pod \"community-operators-c8fzq\" (UID: \"ce15f26e-387c-48a0-86af-1cd73371bc7a\") " pod="openshift-marketplace/community-operators-c8fzq" Dec 10 19:45:33 crc kubenswrapper[4828]: I1210 19:45:33.684935 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce15f26e-387c-48a0-86af-1cd73371bc7a-utilities\") pod \"community-operators-c8fzq\" (UID: \"ce15f26e-387c-48a0-86af-1cd73371bc7a\") " pod="openshift-marketplace/community-operators-c8fzq" Dec 10 19:45:33 crc kubenswrapper[4828]: I1210 19:45:33.684977 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtq9m\" (UniqueName: \"kubernetes.io/projected/ce15f26e-387c-48a0-86af-1cd73371bc7a-kube-api-access-rtq9m\") pod \"community-operators-c8fzq\" (UID: \"ce15f26e-387c-48a0-86af-1cd73371bc7a\") " pod="openshift-marketplace/community-operators-c8fzq" Dec 10 19:45:33 crc kubenswrapper[4828]: I1210 19:45:33.685023 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce15f26e-387c-48a0-86af-1cd73371bc7a-catalog-content\") pod \"community-operators-c8fzq\" (UID: \"ce15f26e-387c-48a0-86af-1cd73371bc7a\") " pod="openshift-marketplace/community-operators-c8fzq" Dec 10 19:45:33 crc kubenswrapper[4828]: I1210 19:45:33.685414 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce15f26e-387c-48a0-86af-1cd73371bc7a-utilities\") pod \"community-operators-c8fzq\" (UID: \"ce15f26e-387c-48a0-86af-1cd73371bc7a\") " pod="openshift-marketplace/community-operators-c8fzq" Dec 10 19:45:33 crc kubenswrapper[4828]: I1210 19:45:33.685598 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce15f26e-387c-48a0-86af-1cd73371bc7a-catalog-content\") pod \"community-operators-c8fzq\" (UID: \"ce15f26e-387c-48a0-86af-1cd73371bc7a\") " pod="openshift-marketplace/community-operators-c8fzq" Dec 10 19:45:33 crc kubenswrapper[4828]: I1210 19:45:33.708086 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtq9m\" (UniqueName: \"kubernetes.io/projected/ce15f26e-387c-48a0-86af-1cd73371bc7a-kube-api-access-rtq9m\") pod \"community-operators-c8fzq\" (UID: \"ce15f26e-387c-48a0-86af-1cd73371bc7a\") " pod="openshift-marketplace/community-operators-c8fzq" Dec 10 19:45:33 crc kubenswrapper[4828]: I1210 19:45:33.822222 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c8fzq" Dec 10 19:45:34 crc kubenswrapper[4828]: I1210 19:45:34.496599 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c8fzq"] Dec 10 19:45:35 crc kubenswrapper[4828]: I1210 19:45:35.117854 4828 generic.go:334] "Generic (PLEG): container finished" podID="ce15f26e-387c-48a0-86af-1cd73371bc7a" containerID="6b44275fa9a4f4490ed9d666125f0e8e563ce441030475bb55189c5d1db0f8d8" exitCode=0 Dec 10 19:45:35 crc kubenswrapper[4828]: I1210 19:45:35.117920 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c8fzq" event={"ID":"ce15f26e-387c-48a0-86af-1cd73371bc7a","Type":"ContainerDied","Data":"6b44275fa9a4f4490ed9d666125f0e8e563ce441030475bb55189c5d1db0f8d8"} Dec 10 19:45:35 crc kubenswrapper[4828]: I1210 19:45:35.118138 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c8fzq" event={"ID":"ce15f26e-387c-48a0-86af-1cd73371bc7a","Type":"ContainerStarted","Data":"584f99b016fab8198623cd1df4cd431a21d8f151f5560ccc80559ed7c33bf11f"} Dec 10 19:45:36 crc kubenswrapper[4828]: I1210 19:45:36.132295 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c8fzq" event={"ID":"ce15f26e-387c-48a0-86af-1cd73371bc7a","Type":"ContainerStarted","Data":"0199d1ec97e852e412771b7e2589d188ea4b159e92138486520e70d5f0925a15"} Dec 10 19:45:37 crc kubenswrapper[4828]: I1210 19:45:37.149262 4828 generic.go:334] "Generic (PLEG): container finished" podID="ce15f26e-387c-48a0-86af-1cd73371bc7a" containerID="0199d1ec97e852e412771b7e2589d188ea4b159e92138486520e70d5f0925a15" exitCode=0 Dec 10 19:45:37 crc kubenswrapper[4828]: I1210 19:45:37.149347 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c8fzq" event={"ID":"ce15f26e-387c-48a0-86af-1cd73371bc7a","Type":"ContainerDied","Data":"0199d1ec97e852e412771b7e2589d188ea4b159e92138486520e70d5f0925a15"} Dec 10 19:45:38 crc kubenswrapper[4828]: I1210 19:45:38.170625 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c8fzq" event={"ID":"ce15f26e-387c-48a0-86af-1cd73371bc7a","Type":"ContainerStarted","Data":"e2c58fa5e1437d5ed9c4331d897459fcecf5695e71196ba884106c5753e29911"} Dec 10 19:45:38 crc kubenswrapper[4828]: I1210 19:45:38.207522 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c8fzq" podStartSLOduration=2.734941707 podStartE2EDuration="5.207488443s" podCreationTimestamp="2025-12-10 19:45:33 +0000 UTC" firstStartedPulling="2025-12-10 19:45:35.120334064 +0000 UTC m=+3015.630945089" lastFinishedPulling="2025-12-10 19:45:37.59288082 +0000 UTC m=+3018.103491825" observedRunningTime="2025-12-10 19:45:38.1982911 +0000 UTC m=+3018.708902125" watchObservedRunningTime="2025-12-10 19:45:38.207488443 +0000 UTC m=+3018.718099448" Dec 10 19:45:39 crc kubenswrapper[4828]: I1210 19:45:39.796623 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:45:39 crc kubenswrapper[4828]: E1210 19:45:39.796898 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:45:42 crc kubenswrapper[4828]: I1210 19:45:42.200132 4828 scope.go:117] "RemoveContainer" containerID="b0db0480ae039902518f264b8731bfb8c9a45bb71ef1a6b105377e32771ef083" Dec 10 19:45:43 crc kubenswrapper[4828]: I1210 19:45:43.822625 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c8fzq" Dec 10 19:45:43 crc kubenswrapper[4828]: I1210 19:45:43.822981 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c8fzq" Dec 10 19:45:43 crc kubenswrapper[4828]: I1210 19:45:43.880455 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c8fzq" Dec 10 19:45:44 crc kubenswrapper[4828]: I1210 19:45:44.283721 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c8fzq" Dec 10 19:45:44 crc kubenswrapper[4828]: I1210 19:45:44.348172 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c8fzq"] Dec 10 19:45:46 crc kubenswrapper[4828]: I1210 19:45:46.250838 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-c8fzq" podUID="ce15f26e-387c-48a0-86af-1cd73371bc7a" containerName="registry-server" containerID="cri-o://e2c58fa5e1437d5ed9c4331d897459fcecf5695e71196ba884106c5753e29911" gracePeriod=2 Dec 10 19:45:46 crc kubenswrapper[4828]: I1210 19:45:46.900748 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c8fzq" Dec 10 19:45:46 crc kubenswrapper[4828]: I1210 19:45:46.937874 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce15f26e-387c-48a0-86af-1cd73371bc7a-catalog-content\") pod \"ce15f26e-387c-48a0-86af-1cd73371bc7a\" (UID: \"ce15f26e-387c-48a0-86af-1cd73371bc7a\") " Dec 10 19:45:46 crc kubenswrapper[4828]: I1210 19:45:46.939911 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtq9m\" (UniqueName: \"kubernetes.io/projected/ce15f26e-387c-48a0-86af-1cd73371bc7a-kube-api-access-rtq9m\") pod \"ce15f26e-387c-48a0-86af-1cd73371bc7a\" (UID: \"ce15f26e-387c-48a0-86af-1cd73371bc7a\") " Dec 10 19:45:46 crc kubenswrapper[4828]: I1210 19:45:46.940057 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce15f26e-387c-48a0-86af-1cd73371bc7a-utilities\") pod \"ce15f26e-387c-48a0-86af-1cd73371bc7a\" (UID: \"ce15f26e-387c-48a0-86af-1cd73371bc7a\") " Dec 10 19:45:46 crc kubenswrapper[4828]: I1210 19:45:46.940730 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce15f26e-387c-48a0-86af-1cd73371bc7a-utilities" (OuterVolumeSpecName: "utilities") pod "ce15f26e-387c-48a0-86af-1cd73371bc7a" (UID: "ce15f26e-387c-48a0-86af-1cd73371bc7a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:45:46 crc kubenswrapper[4828]: I1210 19:45:46.943216 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce15f26e-387c-48a0-86af-1cd73371bc7a-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:45:46 crc kubenswrapper[4828]: I1210 19:45:46.946134 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce15f26e-387c-48a0-86af-1cd73371bc7a-kube-api-access-rtq9m" (OuterVolumeSpecName: "kube-api-access-rtq9m") pod "ce15f26e-387c-48a0-86af-1cd73371bc7a" (UID: "ce15f26e-387c-48a0-86af-1cd73371bc7a"). InnerVolumeSpecName "kube-api-access-rtq9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:45:46 crc kubenswrapper[4828]: I1210 19:45:46.999865 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce15f26e-387c-48a0-86af-1cd73371bc7a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce15f26e-387c-48a0-86af-1cd73371bc7a" (UID: "ce15f26e-387c-48a0-86af-1cd73371bc7a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.045743 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce15f26e-387c-48a0-86af-1cd73371bc7a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.045779 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtq9m\" (UniqueName: \"kubernetes.io/projected/ce15f26e-387c-48a0-86af-1cd73371bc7a-kube-api-access-rtq9m\") on node \"crc\" DevicePath \"\"" Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.264251 4828 generic.go:334] "Generic (PLEG): container finished" podID="ce15f26e-387c-48a0-86af-1cd73371bc7a" containerID="e2c58fa5e1437d5ed9c4331d897459fcecf5695e71196ba884106c5753e29911" exitCode=0 Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.264291 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c8fzq" event={"ID":"ce15f26e-387c-48a0-86af-1cd73371bc7a","Type":"ContainerDied","Data":"e2c58fa5e1437d5ed9c4331d897459fcecf5695e71196ba884106c5753e29911"} Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.264315 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c8fzq" event={"ID":"ce15f26e-387c-48a0-86af-1cd73371bc7a","Type":"ContainerDied","Data":"584f99b016fab8198623cd1df4cd431a21d8f151f5560ccc80559ed7c33bf11f"} Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.264331 4828 scope.go:117] "RemoveContainer" containerID="e2c58fa5e1437d5ed9c4331d897459fcecf5695e71196ba884106c5753e29911" Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.264331 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c8fzq" Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.290694 4828 scope.go:117] "RemoveContainer" containerID="0199d1ec97e852e412771b7e2589d188ea4b159e92138486520e70d5f0925a15" Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.305007 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-c8fzq"] Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.317413 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-c8fzq"] Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.332433 4828 scope.go:117] "RemoveContainer" containerID="6b44275fa9a4f4490ed9d666125f0e8e563ce441030475bb55189c5d1db0f8d8" Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.370383 4828 scope.go:117] "RemoveContainer" containerID="e2c58fa5e1437d5ed9c4331d897459fcecf5695e71196ba884106c5753e29911" Dec 10 19:45:47 crc kubenswrapper[4828]: E1210 19:45:47.371005 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2c58fa5e1437d5ed9c4331d897459fcecf5695e71196ba884106c5753e29911\": container with ID starting with e2c58fa5e1437d5ed9c4331d897459fcecf5695e71196ba884106c5753e29911 not found: ID does not exist" containerID="e2c58fa5e1437d5ed9c4331d897459fcecf5695e71196ba884106c5753e29911" Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.371169 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2c58fa5e1437d5ed9c4331d897459fcecf5695e71196ba884106c5753e29911"} err="failed to get container status \"e2c58fa5e1437d5ed9c4331d897459fcecf5695e71196ba884106c5753e29911\": rpc error: code = NotFound desc = could not find container \"e2c58fa5e1437d5ed9c4331d897459fcecf5695e71196ba884106c5753e29911\": container with ID starting with e2c58fa5e1437d5ed9c4331d897459fcecf5695e71196ba884106c5753e29911 not found: ID does not exist" Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.371279 4828 scope.go:117] "RemoveContainer" containerID="0199d1ec97e852e412771b7e2589d188ea4b159e92138486520e70d5f0925a15" Dec 10 19:45:47 crc kubenswrapper[4828]: E1210 19:45:47.371710 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0199d1ec97e852e412771b7e2589d188ea4b159e92138486520e70d5f0925a15\": container with ID starting with 0199d1ec97e852e412771b7e2589d188ea4b159e92138486520e70d5f0925a15 not found: ID does not exist" containerID="0199d1ec97e852e412771b7e2589d188ea4b159e92138486520e70d5f0925a15" Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.371753 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0199d1ec97e852e412771b7e2589d188ea4b159e92138486520e70d5f0925a15"} err="failed to get container status \"0199d1ec97e852e412771b7e2589d188ea4b159e92138486520e70d5f0925a15\": rpc error: code = NotFound desc = could not find container \"0199d1ec97e852e412771b7e2589d188ea4b159e92138486520e70d5f0925a15\": container with ID starting with 0199d1ec97e852e412771b7e2589d188ea4b159e92138486520e70d5f0925a15 not found: ID does not exist" Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.371781 4828 scope.go:117] "RemoveContainer" containerID="6b44275fa9a4f4490ed9d666125f0e8e563ce441030475bb55189c5d1db0f8d8" Dec 10 19:45:47 crc kubenswrapper[4828]: E1210 19:45:47.372372 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b44275fa9a4f4490ed9d666125f0e8e563ce441030475bb55189c5d1db0f8d8\": container with ID starting with 6b44275fa9a4f4490ed9d666125f0e8e563ce441030475bb55189c5d1db0f8d8 not found: ID does not exist" containerID="6b44275fa9a4f4490ed9d666125f0e8e563ce441030475bb55189c5d1db0f8d8" Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.372415 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b44275fa9a4f4490ed9d666125f0e8e563ce441030475bb55189c5d1db0f8d8"} err="failed to get container status \"6b44275fa9a4f4490ed9d666125f0e8e563ce441030475bb55189c5d1db0f8d8\": rpc error: code = NotFound desc = could not find container \"6b44275fa9a4f4490ed9d666125f0e8e563ce441030475bb55189c5d1db0f8d8\": container with ID starting with 6b44275fa9a4f4490ed9d666125f0e8e563ce441030475bb55189c5d1db0f8d8 not found: ID does not exist" Dec 10 19:45:47 crc kubenswrapper[4828]: I1210 19:45:47.801639 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce15f26e-387c-48a0-86af-1cd73371bc7a" path="/var/lib/kubelet/pods/ce15f26e-387c-48a0-86af-1cd73371bc7a/volumes" Dec 10 19:45:50 crc kubenswrapper[4828]: I1210 19:45:50.789283 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:45:50 crc kubenswrapper[4828]: E1210 19:45:50.790235 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:46:03 crc kubenswrapper[4828]: I1210 19:46:03.791693 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:46:03 crc kubenswrapper[4828]: E1210 19:46:03.792574 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:46:17 crc kubenswrapper[4828]: I1210 19:46:17.789698 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:46:17 crc kubenswrapper[4828]: E1210 19:46:17.790627 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:46:29 crc kubenswrapper[4828]: I1210 19:46:29.795982 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:46:29 crc kubenswrapper[4828]: E1210 19:46:29.796662 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:46:43 crc kubenswrapper[4828]: I1210 19:46:43.789347 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:46:43 crc kubenswrapper[4828]: E1210 19:46:43.790087 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:46:57 crc kubenswrapper[4828]: I1210 19:46:57.789205 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:46:57 crc kubenswrapper[4828]: E1210 19:46:57.790135 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:47:02 crc kubenswrapper[4828]: I1210 19:47:02.067088 4828 generic.go:334] "Generic (PLEG): container finished" podID="f5034a0e-652b-414a-a36f-e7f4f34da750" containerID="346abaccfc8d2cdbd3379f61f60e5bedca51d1707b7c0d22103d357d4034d306" exitCode=0 Dec 10 19:47:02 crc kubenswrapper[4828]: I1210 19:47:02.067172 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" event={"ID":"f5034a0e-652b-414a-a36f-e7f4f34da750","Type":"ContainerDied","Data":"346abaccfc8d2cdbd3379f61f60e5bedca51d1707b7c0d22103d357d4034d306"} Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.524928 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.672325 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-1\") pod \"f5034a0e-652b-414a-a36f-e7f4f34da750\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.672375 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-inventory\") pod \"f5034a0e-652b-414a-a36f-e7f4f34da750\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.672546 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-2\") pod \"f5034a0e-652b-414a-a36f-e7f4f34da750\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.672580 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ssh-key\") pod \"f5034a0e-652b-414a-a36f-e7f4f34da750\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.672614 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgbt2\" (UniqueName: \"kubernetes.io/projected/f5034a0e-652b-414a-a36f-e7f4f34da750-kube-api-access-kgbt2\") pod \"f5034a0e-652b-414a-a36f-e7f4f34da750\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.672693 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-0\") pod \"f5034a0e-652b-414a-a36f-e7f4f34da750\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.672763 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-telemetry-combined-ca-bundle\") pod \"f5034a0e-652b-414a-a36f-e7f4f34da750\" (UID: \"f5034a0e-652b-414a-a36f-e7f4f34da750\") " Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.678091 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "f5034a0e-652b-414a-a36f-e7f4f34da750" (UID: "f5034a0e-652b-414a-a36f-e7f4f34da750"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.679694 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5034a0e-652b-414a-a36f-e7f4f34da750-kube-api-access-kgbt2" (OuterVolumeSpecName: "kube-api-access-kgbt2") pod "f5034a0e-652b-414a-a36f-e7f4f34da750" (UID: "f5034a0e-652b-414a-a36f-e7f4f34da750"). InnerVolumeSpecName "kube-api-access-kgbt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.726590 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "f5034a0e-652b-414a-a36f-e7f4f34da750" (UID: "f5034a0e-652b-414a-a36f-e7f4f34da750"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.727017 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-inventory" (OuterVolumeSpecName: "inventory") pod "f5034a0e-652b-414a-a36f-e7f4f34da750" (UID: "f5034a0e-652b-414a-a36f-e7f4f34da750"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.730482 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "f5034a0e-652b-414a-a36f-e7f4f34da750" (UID: "f5034a0e-652b-414a-a36f-e7f4f34da750"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.739251 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "f5034a0e-652b-414a-a36f-e7f4f34da750" (UID: "f5034a0e-652b-414a-a36f-e7f4f34da750"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.742525 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f5034a0e-652b-414a-a36f-e7f4f34da750" (UID: "f5034a0e-652b-414a-a36f-e7f4f34da750"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.775193 4828 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.775247 4828 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.775264 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.775278 4828 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.775291 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.775302 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgbt2\" (UniqueName: \"kubernetes.io/projected/f5034a0e-652b-414a-a36f-e7f4f34da750-kube-api-access-kgbt2\") on node \"crc\" DevicePath \"\"" Dec 10 19:47:03 crc kubenswrapper[4828]: I1210 19:47:03.775314 4828 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f5034a0e-652b-414a-a36f-e7f4f34da750-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.089596 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" event={"ID":"f5034a0e-652b-414a-a36f-e7f4f34da750","Type":"ContainerDied","Data":"2f833b94c3ffdfb3aeb1272c4ecfd9a3cfcd9d7e463ac33fe059eed3c54b81cd"} Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.089638 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f833b94c3ffdfb3aeb1272c4ecfd9a3cfcd9d7e463ac33fe059eed3c54b81cd" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.089663 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-csrqs" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.187035 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w"] Dec 10 19:47:04 crc kubenswrapper[4828]: E1210 19:47:04.189157 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce15f26e-387c-48a0-86af-1cd73371bc7a" containerName="extract-content" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.189187 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce15f26e-387c-48a0-86af-1cd73371bc7a" containerName="extract-content" Dec 10 19:47:04 crc kubenswrapper[4828]: E1210 19:47:04.189228 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5034a0e-652b-414a-a36f-e7f4f34da750" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.189241 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5034a0e-652b-414a-a36f-e7f4f34da750" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 10 19:47:04 crc kubenswrapper[4828]: E1210 19:47:04.189254 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce15f26e-387c-48a0-86af-1cd73371bc7a" containerName="extract-utilities" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.189262 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce15f26e-387c-48a0-86af-1cd73371bc7a" containerName="extract-utilities" Dec 10 19:47:04 crc kubenswrapper[4828]: E1210 19:47:04.189276 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce15f26e-387c-48a0-86af-1cd73371bc7a" containerName="registry-server" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.189283 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce15f26e-387c-48a0-86af-1cd73371bc7a" containerName="registry-server" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.189586 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5034a0e-652b-414a-a36f-e7f4f34da750" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.189627 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce15f26e-387c-48a0-86af-1cd73371bc7a" containerName="registry-server" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.190674 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.192960 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.193858 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.195186 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.195200 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.204662 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.208820 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w"] Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.286071 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.286128 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.286190 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.286241 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.286287 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.286322 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwjhp\" (UniqueName: \"kubernetes.io/projected/68109455-5463-42d8-9ef5-d2c667132032-kube-api-access-hwjhp\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.286443 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.388031 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.388090 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.388128 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.388174 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.388213 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.388245 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwjhp\" (UniqueName: \"kubernetes.io/projected/68109455-5463-42d8-9ef5-d2c667132032-kube-api-access-hwjhp\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.388900 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.393401 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.393789 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.394552 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.394942 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.398441 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.399326 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.404711 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwjhp\" (UniqueName: \"kubernetes.io/projected/68109455-5463-42d8-9ef5-d2c667132032-kube-api-access-hwjhp\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:04 crc kubenswrapper[4828]: I1210 19:47:04.513671 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:47:05 crc kubenswrapper[4828]: I1210 19:47:05.047514 4828 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:47:05 crc kubenswrapper[4828]: I1210 19:47:05.047624 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w"] Dec 10 19:47:05 crc kubenswrapper[4828]: I1210 19:47:05.101790 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" event={"ID":"68109455-5463-42d8-9ef5-d2c667132032","Type":"ContainerStarted","Data":"67f1b083396554bf1f9b9ef29247dbeabb74445a9a00e01dc9cd351b2462a6e5"} Dec 10 19:47:06 crc kubenswrapper[4828]: I1210 19:47:06.113137 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" event={"ID":"68109455-5463-42d8-9ef5-d2c667132032","Type":"ContainerStarted","Data":"b3aa876df4c0ebfc50ec4bbcd9819308a5db98e7a259348274ed9e67adc9ac43"} Dec 10 19:47:06 crc kubenswrapper[4828]: I1210 19:47:06.141561 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" podStartSLOduration=1.498194479 podStartE2EDuration="2.141540293s" podCreationTimestamp="2025-12-10 19:47:04 +0000 UTC" firstStartedPulling="2025-12-10 19:47:05.047219179 +0000 UTC m=+3105.557830184" lastFinishedPulling="2025-12-10 19:47:05.690564993 +0000 UTC m=+3106.201175998" observedRunningTime="2025-12-10 19:47:06.129174166 +0000 UTC m=+3106.639785181" watchObservedRunningTime="2025-12-10 19:47:06.141540293 +0000 UTC m=+3106.652151308" Dec 10 19:47:11 crc kubenswrapper[4828]: I1210 19:47:11.808554 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:47:11 crc kubenswrapper[4828]: E1210 19:47:11.809342 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:47:25 crc kubenswrapper[4828]: I1210 19:47:25.790844 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:47:25 crc kubenswrapper[4828]: E1210 19:47:25.791708 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:47:40 crc kubenswrapper[4828]: I1210 19:47:40.789110 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:47:40 crc kubenswrapper[4828]: E1210 19:47:40.790225 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:47:52 crc kubenswrapper[4828]: I1210 19:47:52.789496 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:47:52 crc kubenswrapper[4828]: E1210 19:47:52.790759 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:48:04 crc kubenswrapper[4828]: I1210 19:48:04.789364 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:48:04 crc kubenswrapper[4828]: E1210 19:48:04.790136 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:48:16 crc kubenswrapper[4828]: I1210 19:48:16.789734 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:48:16 crc kubenswrapper[4828]: E1210 19:48:16.790780 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:48:29 crc kubenswrapper[4828]: I1210 19:48:29.805665 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:48:29 crc kubenswrapper[4828]: E1210 19:48:29.806875 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:48:41 crc kubenswrapper[4828]: I1210 19:48:41.789698 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:48:41 crc kubenswrapper[4828]: E1210 19:48:41.791633 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:48:52 crc kubenswrapper[4828]: I1210 19:48:52.789361 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:48:53 crc kubenswrapper[4828]: I1210 19:48:53.281998 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"558207227d2901f2b3b87b61b1e8400a08f602aa55d126b771706aa1bef786c1"} Dec 10 19:49:00 crc kubenswrapper[4828]: I1210 19:49:00.361143 4828 generic.go:334] "Generic (PLEG): container finished" podID="68109455-5463-42d8-9ef5-d2c667132032" containerID="b3aa876df4c0ebfc50ec4bbcd9819308a5db98e7a259348274ed9e67adc9ac43" exitCode=0 Dec 10 19:49:00 crc kubenswrapper[4828]: I1210 19:49:00.361231 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" event={"ID":"68109455-5463-42d8-9ef5-d2c667132032","Type":"ContainerDied","Data":"b3aa876df4c0ebfc50ec4bbcd9819308a5db98e7a259348274ed9e67adc9ac43"} Dec 10 19:49:01 crc kubenswrapper[4828]: I1210 19:49:01.882416 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:49:01 crc kubenswrapper[4828]: I1210 19:49:01.988634 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-2\") pod \"68109455-5463-42d8-9ef5-d2c667132032\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " Dec 10 19:49:01 crc kubenswrapper[4828]: I1210 19:49:01.988699 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ssh-key\") pod \"68109455-5463-42d8-9ef5-d2c667132032\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " Dec 10 19:49:01 crc kubenswrapper[4828]: I1210 19:49:01.988756 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-0\") pod \"68109455-5463-42d8-9ef5-d2c667132032\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " Dec 10 19:49:01 crc kubenswrapper[4828]: I1210 19:49:01.988859 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwjhp\" (UniqueName: \"kubernetes.io/projected/68109455-5463-42d8-9ef5-d2c667132032-kube-api-access-hwjhp\") pod \"68109455-5463-42d8-9ef5-d2c667132032\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " Dec 10 19:49:01 crc kubenswrapper[4828]: I1210 19:49:01.988984 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-telemetry-power-monitoring-combined-ca-bundle\") pod \"68109455-5463-42d8-9ef5-d2c667132032\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " Dec 10 19:49:01 crc kubenswrapper[4828]: I1210 19:49:01.989029 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-1\") pod \"68109455-5463-42d8-9ef5-d2c667132032\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " Dec 10 19:49:01 crc kubenswrapper[4828]: I1210 19:49:01.989085 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-inventory\") pod \"68109455-5463-42d8-9ef5-d2c667132032\" (UID: \"68109455-5463-42d8-9ef5-d2c667132032\") " Dec 10 19:49:01 crc kubenswrapper[4828]: I1210 19:49:01.995993 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "68109455-5463-42d8-9ef5-d2c667132032" (UID: "68109455-5463-42d8-9ef5-d2c667132032"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.012074 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68109455-5463-42d8-9ef5-d2c667132032-kube-api-access-hwjhp" (OuterVolumeSpecName: "kube-api-access-hwjhp") pod "68109455-5463-42d8-9ef5-d2c667132032" (UID: "68109455-5463-42d8-9ef5-d2c667132032"). InnerVolumeSpecName "kube-api-access-hwjhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.026032 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "68109455-5463-42d8-9ef5-d2c667132032" (UID: "68109455-5463-42d8-9ef5-d2c667132032"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.026065 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "68109455-5463-42d8-9ef5-d2c667132032" (UID: "68109455-5463-42d8-9ef5-d2c667132032"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.026170 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "68109455-5463-42d8-9ef5-d2c667132032" (UID: "68109455-5463-42d8-9ef5-d2c667132032"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.026346 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "68109455-5463-42d8-9ef5-d2c667132032" (UID: "68109455-5463-42d8-9ef5-d2c667132032"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.039134 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-inventory" (OuterVolumeSpecName: "inventory") pod "68109455-5463-42d8-9ef5-d2c667132032" (UID: "68109455-5463-42d8-9ef5-d2c667132032"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.092560 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.092598 4828 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.092611 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.092620 4828 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.092630 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwjhp\" (UniqueName: \"kubernetes.io/projected/68109455-5463-42d8-9ef5-d2c667132032-kube-api-access-hwjhp\") on node \"crc\" DevicePath \"\"" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.092639 4828 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.092653 4828 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/68109455-5463-42d8-9ef5-d2c667132032-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.386044 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" event={"ID":"68109455-5463-42d8-9ef5-d2c667132032","Type":"ContainerDied","Data":"67f1b083396554bf1f9b9ef29247dbeabb74445a9a00e01dc9cd351b2462a6e5"} Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.386500 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67f1b083396554bf1f9b9ef29247dbeabb74445a9a00e01dc9cd351b2462a6e5" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.386151 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.529640 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42"] Dec 10 19:49:02 crc kubenswrapper[4828]: E1210 19:49:02.530362 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68109455-5463-42d8-9ef5-d2c667132032" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.530380 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="68109455-5463-42d8-9ef5-d2c667132032" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.530695 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="68109455-5463-42d8-9ef5-d2c667132032" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.532132 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.535153 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.535402 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.535461 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.535508 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qpslc" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.535789 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.542187 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42"] Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.707664 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxjns\" (UniqueName: \"kubernetes.io/projected/94949893-8bdc-4a05-b105-8759681782ae-kube-api-access-lxjns\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2lb42\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.708054 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2lb42\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.708240 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2lb42\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.708453 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2lb42\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.708572 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2lb42\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.810126 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxjns\" (UniqueName: \"kubernetes.io/projected/94949893-8bdc-4a05-b105-8759681782ae-kube-api-access-lxjns\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2lb42\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.810201 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2lb42\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.810269 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2lb42\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.810389 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2lb42\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.810414 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2lb42\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.815683 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2lb42\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.816151 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2lb42\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.816592 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2lb42\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.816893 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2lb42\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.828830 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxjns\" (UniqueName: \"kubernetes.io/projected/94949893-8bdc-4a05-b105-8759681782ae-kube-api-access-lxjns\") pod \"logging-edpm-deployment-openstack-edpm-ipam-2lb42\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:02 crc kubenswrapper[4828]: I1210 19:49:02.864817 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:03 crc kubenswrapper[4828]: I1210 19:49:03.427445 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42"] Dec 10 19:49:04 crc kubenswrapper[4828]: I1210 19:49:04.407007 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" event={"ID":"94949893-8bdc-4a05-b105-8759681782ae","Type":"ContainerStarted","Data":"074af227e5d75202ad30c44eb6be3e28e6777510307b3a8a75a1d516ba27c32f"} Dec 10 19:49:04 crc kubenswrapper[4828]: I1210 19:49:04.407361 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" event={"ID":"94949893-8bdc-4a05-b105-8759681782ae","Type":"ContainerStarted","Data":"e53d95f3cdb777f30f6ea3b1b32426689d2cf38e776aab50dd553a72b710c0e3"} Dec 10 19:49:04 crc kubenswrapper[4828]: I1210 19:49:04.429644 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" podStartSLOduration=1.9027497169999998 podStartE2EDuration="2.429627317s" podCreationTimestamp="2025-12-10 19:49:02 +0000 UTC" firstStartedPulling="2025-12-10 19:49:03.432083255 +0000 UTC m=+3223.942694260" lastFinishedPulling="2025-12-10 19:49:03.958960855 +0000 UTC m=+3224.469571860" observedRunningTime="2025-12-10 19:49:04.425125338 +0000 UTC m=+3224.935736343" watchObservedRunningTime="2025-12-10 19:49:04.429627317 +0000 UTC m=+3224.940238322" Dec 10 19:49:19 crc kubenswrapper[4828]: E1210 19:49:19.541656 4828 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94949893_8bdc_4a05_b105_8759681782ae.slice/crio-074af227e5d75202ad30c44eb6be3e28e6777510307b3a8a75a1d516ba27c32f.scope\": RecentStats: unable to find data in memory cache]" Dec 10 19:49:19 crc kubenswrapper[4828]: I1210 19:49:19.569623 4828 generic.go:334] "Generic (PLEG): container finished" podID="94949893-8bdc-4a05-b105-8759681782ae" containerID="074af227e5d75202ad30c44eb6be3e28e6777510307b3a8a75a1d516ba27c32f" exitCode=0 Dec 10 19:49:19 crc kubenswrapper[4828]: I1210 19:49:19.569669 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" event={"ID":"94949893-8bdc-4a05-b105-8759681782ae","Type":"ContainerDied","Data":"074af227e5d75202ad30c44eb6be3e28e6777510307b3a8a75a1d516ba27c32f"} Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.017042 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.107082 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxjns\" (UniqueName: \"kubernetes.io/projected/94949893-8bdc-4a05-b105-8759681782ae-kube-api-access-lxjns\") pod \"94949893-8bdc-4a05-b105-8759681782ae\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.107153 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-inventory\") pod \"94949893-8bdc-4a05-b105-8759681782ae\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.107184 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-logging-compute-config-data-0\") pod \"94949893-8bdc-4a05-b105-8759681782ae\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.107451 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-logging-compute-config-data-1\") pod \"94949893-8bdc-4a05-b105-8759681782ae\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.107526 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-ssh-key\") pod \"94949893-8bdc-4a05-b105-8759681782ae\" (UID: \"94949893-8bdc-4a05-b105-8759681782ae\") " Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.113494 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94949893-8bdc-4a05-b105-8759681782ae-kube-api-access-lxjns" (OuterVolumeSpecName: "kube-api-access-lxjns") pod "94949893-8bdc-4a05-b105-8759681782ae" (UID: "94949893-8bdc-4a05-b105-8759681782ae"). InnerVolumeSpecName "kube-api-access-lxjns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.140608 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "94949893-8bdc-4a05-b105-8759681782ae" (UID: "94949893-8bdc-4a05-b105-8759681782ae"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.141033 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "94949893-8bdc-4a05-b105-8759681782ae" (UID: "94949893-8bdc-4a05-b105-8759681782ae"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.156994 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-inventory" (OuterVolumeSpecName: "inventory") pod "94949893-8bdc-4a05-b105-8759681782ae" (UID: "94949893-8bdc-4a05-b105-8759681782ae"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.165744 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "94949893-8bdc-4a05-b105-8759681782ae" (UID: "94949893-8bdc-4a05-b105-8759681782ae"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.209758 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxjns\" (UniqueName: \"kubernetes.io/projected/94949893-8bdc-4a05-b105-8759681782ae-kube-api-access-lxjns\") on node \"crc\" DevicePath \"\"" Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.209828 4828 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-inventory\") on node \"crc\" DevicePath \"\"" Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.209841 4828 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.209851 4828 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.209860 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94949893-8bdc-4a05-b105-8759681782ae-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.591556 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" event={"ID":"94949893-8bdc-4a05-b105-8759681782ae","Type":"ContainerDied","Data":"e53d95f3cdb777f30f6ea3b1b32426689d2cf38e776aab50dd553a72b710c0e3"} Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.591851 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e53d95f3cdb777f30f6ea3b1b32426689d2cf38e776aab50dd553a72b710c0e3" Dec 10 19:49:21 crc kubenswrapper[4828]: I1210 19:49:21.591634 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-2lb42" Dec 10 19:51:21 crc kubenswrapper[4828]: I1210 19:51:21.230338 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:51:21 crc kubenswrapper[4828]: I1210 19:51:21.230897 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:51:51 crc kubenswrapper[4828]: I1210 19:51:51.230703 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:51:51 crc kubenswrapper[4828]: I1210 19:51:51.231341 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:52:21 crc kubenswrapper[4828]: I1210 19:52:21.230232 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:52:21 crc kubenswrapper[4828]: I1210 19:52:21.230835 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:52:21 crc kubenswrapper[4828]: I1210 19:52:21.230890 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 19:52:21 crc kubenswrapper[4828]: I1210 19:52:21.231688 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"558207227d2901f2b3b87b61b1e8400a08f602aa55d126b771706aa1bef786c1"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:52:21 crc kubenswrapper[4828]: I1210 19:52:21.231739 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://558207227d2901f2b3b87b61b1e8400a08f602aa55d126b771706aa1bef786c1" gracePeriod=600 Dec 10 19:52:21 crc kubenswrapper[4828]: I1210 19:52:21.674626 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="558207227d2901f2b3b87b61b1e8400a08f602aa55d126b771706aa1bef786c1" exitCode=0 Dec 10 19:52:21 crc kubenswrapper[4828]: I1210 19:52:21.674698 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"558207227d2901f2b3b87b61b1e8400a08f602aa55d126b771706aa1bef786c1"} Dec 10 19:52:21 crc kubenswrapper[4828]: I1210 19:52:21.675017 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd"} Dec 10 19:52:21 crc kubenswrapper[4828]: I1210 19:52:21.675041 4828 scope.go:117] "RemoveContainer" containerID="d08e3727099e9f275def5d3feb40649df40a638ab5650de7e60a54d90c0a1a83" Dec 10 19:52:51 crc kubenswrapper[4828]: I1210 19:52:51.787544 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8k7z5"] Dec 10 19:52:51 crc kubenswrapper[4828]: E1210 19:52:51.788842 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94949893-8bdc-4a05-b105-8759681782ae" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 10 19:52:51 crc kubenswrapper[4828]: I1210 19:52:51.788858 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="94949893-8bdc-4a05-b105-8759681782ae" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 10 19:52:51 crc kubenswrapper[4828]: I1210 19:52:51.789131 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="94949893-8bdc-4a05-b105-8759681782ae" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 10 19:52:51 crc kubenswrapper[4828]: I1210 19:52:51.794452 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8k7z5" Dec 10 19:52:51 crc kubenswrapper[4828]: I1210 19:52:51.810630 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8k7z5"] Dec 10 19:52:51 crc kubenswrapper[4828]: I1210 19:52:51.949384 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1776efd4-52c9-4579-aa7f-e14486935382-utilities\") pod \"redhat-operators-8k7z5\" (UID: \"1776efd4-52c9-4579-aa7f-e14486935382\") " pod="openshift-marketplace/redhat-operators-8k7z5" Dec 10 19:52:51 crc kubenswrapper[4828]: I1210 19:52:51.949866 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1776efd4-52c9-4579-aa7f-e14486935382-catalog-content\") pod \"redhat-operators-8k7z5\" (UID: \"1776efd4-52c9-4579-aa7f-e14486935382\") " pod="openshift-marketplace/redhat-operators-8k7z5" Dec 10 19:52:51 crc kubenswrapper[4828]: I1210 19:52:51.950232 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvwh8\" (UniqueName: \"kubernetes.io/projected/1776efd4-52c9-4579-aa7f-e14486935382-kube-api-access-hvwh8\") pod \"redhat-operators-8k7z5\" (UID: \"1776efd4-52c9-4579-aa7f-e14486935382\") " pod="openshift-marketplace/redhat-operators-8k7z5" Dec 10 19:52:52 crc kubenswrapper[4828]: I1210 19:52:52.054231 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1776efd4-52c9-4579-aa7f-e14486935382-catalog-content\") pod \"redhat-operators-8k7z5\" (UID: \"1776efd4-52c9-4579-aa7f-e14486935382\") " pod="openshift-marketplace/redhat-operators-8k7z5" Dec 10 19:52:52 crc kubenswrapper[4828]: I1210 19:52:52.054464 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvwh8\" (UniqueName: \"kubernetes.io/projected/1776efd4-52c9-4579-aa7f-e14486935382-kube-api-access-hvwh8\") pod \"redhat-operators-8k7z5\" (UID: \"1776efd4-52c9-4579-aa7f-e14486935382\") " pod="openshift-marketplace/redhat-operators-8k7z5" Dec 10 19:52:52 crc kubenswrapper[4828]: I1210 19:52:52.054734 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1776efd4-52c9-4579-aa7f-e14486935382-utilities\") pod \"redhat-operators-8k7z5\" (UID: \"1776efd4-52c9-4579-aa7f-e14486935382\") " pod="openshift-marketplace/redhat-operators-8k7z5" Dec 10 19:52:52 crc kubenswrapper[4828]: I1210 19:52:52.054886 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1776efd4-52c9-4579-aa7f-e14486935382-catalog-content\") pod \"redhat-operators-8k7z5\" (UID: \"1776efd4-52c9-4579-aa7f-e14486935382\") " pod="openshift-marketplace/redhat-operators-8k7z5" Dec 10 19:52:52 crc kubenswrapper[4828]: I1210 19:52:52.055299 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1776efd4-52c9-4579-aa7f-e14486935382-utilities\") pod \"redhat-operators-8k7z5\" (UID: \"1776efd4-52c9-4579-aa7f-e14486935382\") " pod="openshift-marketplace/redhat-operators-8k7z5" Dec 10 19:52:52 crc kubenswrapper[4828]: I1210 19:52:52.078624 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvwh8\" (UniqueName: \"kubernetes.io/projected/1776efd4-52c9-4579-aa7f-e14486935382-kube-api-access-hvwh8\") pod \"redhat-operators-8k7z5\" (UID: \"1776efd4-52c9-4579-aa7f-e14486935382\") " pod="openshift-marketplace/redhat-operators-8k7z5" Dec 10 19:52:52 crc kubenswrapper[4828]: I1210 19:52:52.124969 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8k7z5" Dec 10 19:52:52 crc kubenswrapper[4828]: I1210 19:52:52.619434 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8k7z5"] Dec 10 19:52:53 crc kubenswrapper[4828]: I1210 19:52:53.018832 4828 generic.go:334] "Generic (PLEG): container finished" podID="1776efd4-52c9-4579-aa7f-e14486935382" containerID="763c9b3ddb8ad3ea3ab1bf33ea297c216c60a5a581cd587139febbd215b4df82" exitCode=0 Dec 10 19:52:53 crc kubenswrapper[4828]: I1210 19:52:53.018888 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8k7z5" event={"ID":"1776efd4-52c9-4579-aa7f-e14486935382","Type":"ContainerDied","Data":"763c9b3ddb8ad3ea3ab1bf33ea297c216c60a5a581cd587139febbd215b4df82"} Dec 10 19:52:53 crc kubenswrapper[4828]: I1210 19:52:53.018913 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8k7z5" event={"ID":"1776efd4-52c9-4579-aa7f-e14486935382","Type":"ContainerStarted","Data":"5ca375be6afc7d0f7e73db00fdb3f09b3ed7a5f26f9fcb3f4d80880b91defef6"} Dec 10 19:52:53 crc kubenswrapper[4828]: I1210 19:52:53.021257 4828 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 19:52:54 crc kubenswrapper[4828]: I1210 19:52:54.032711 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8k7z5" event={"ID":"1776efd4-52c9-4579-aa7f-e14486935382","Type":"ContainerStarted","Data":"af2c903e8d75ef0dc2c3ff6bb45ed1a1802ded842511513832e24a91fddd0696"} Dec 10 19:52:57 crc kubenswrapper[4828]: I1210 19:52:57.067832 4828 generic.go:334] "Generic (PLEG): container finished" podID="1776efd4-52c9-4579-aa7f-e14486935382" containerID="af2c903e8d75ef0dc2c3ff6bb45ed1a1802ded842511513832e24a91fddd0696" exitCode=0 Dec 10 19:52:57 crc kubenswrapper[4828]: I1210 19:52:57.067840 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8k7z5" event={"ID":"1776efd4-52c9-4579-aa7f-e14486935382","Type":"ContainerDied","Data":"af2c903e8d75ef0dc2c3ff6bb45ed1a1802ded842511513832e24a91fddd0696"} Dec 10 19:52:58 crc kubenswrapper[4828]: I1210 19:52:58.104107 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8k7z5" event={"ID":"1776efd4-52c9-4579-aa7f-e14486935382","Type":"ContainerStarted","Data":"6da6e2fec5c4ddce8e42f1e7c252f4b9fc3fb11ea86a96b17da67780118fd51a"} Dec 10 19:52:58 crc kubenswrapper[4828]: I1210 19:52:58.161113 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8k7z5" podStartSLOduration=2.367660005 podStartE2EDuration="7.161088792s" podCreationTimestamp="2025-12-10 19:52:51 +0000 UTC" firstStartedPulling="2025-12-10 19:52:53.021051351 +0000 UTC m=+3453.531662356" lastFinishedPulling="2025-12-10 19:52:57.814480138 +0000 UTC m=+3458.325091143" observedRunningTime="2025-12-10 19:52:58.13229016 +0000 UTC m=+3458.642901165" watchObservedRunningTime="2025-12-10 19:52:58.161088792 +0000 UTC m=+3458.671699797" Dec 10 19:53:02 crc kubenswrapper[4828]: I1210 19:53:02.125895 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8k7z5" Dec 10 19:53:02 crc kubenswrapper[4828]: I1210 19:53:02.126580 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8k7z5" Dec 10 19:53:03 crc kubenswrapper[4828]: I1210 19:53:03.175994 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8k7z5" podUID="1776efd4-52c9-4579-aa7f-e14486935382" containerName="registry-server" probeResult="failure" output=< Dec 10 19:53:03 crc kubenswrapper[4828]: timeout: failed to connect service ":50051" within 1s Dec 10 19:53:03 crc kubenswrapper[4828]: > Dec 10 19:53:12 crc kubenswrapper[4828]: I1210 19:53:12.183402 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8k7z5" Dec 10 19:53:12 crc kubenswrapper[4828]: I1210 19:53:12.242167 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8k7z5" Dec 10 19:53:12 crc kubenswrapper[4828]: I1210 19:53:12.429188 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8k7z5"] Dec 10 19:53:13 crc kubenswrapper[4828]: I1210 19:53:13.257903 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8k7z5" podUID="1776efd4-52c9-4579-aa7f-e14486935382" containerName="registry-server" containerID="cri-o://6da6e2fec5c4ddce8e42f1e7c252f4b9fc3fb11ea86a96b17da67780118fd51a" gracePeriod=2 Dec 10 19:53:13 crc kubenswrapper[4828]: I1210 19:53:13.839559 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8k7z5" Dec 10 19:53:13 crc kubenswrapper[4828]: I1210 19:53:13.987382 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1776efd4-52c9-4579-aa7f-e14486935382-catalog-content\") pod \"1776efd4-52c9-4579-aa7f-e14486935382\" (UID: \"1776efd4-52c9-4579-aa7f-e14486935382\") " Dec 10 19:53:13 crc kubenswrapper[4828]: I1210 19:53:13.987774 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1776efd4-52c9-4579-aa7f-e14486935382-utilities\") pod \"1776efd4-52c9-4579-aa7f-e14486935382\" (UID: \"1776efd4-52c9-4579-aa7f-e14486935382\") " Dec 10 19:53:13 crc kubenswrapper[4828]: I1210 19:53:13.987953 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvwh8\" (UniqueName: \"kubernetes.io/projected/1776efd4-52c9-4579-aa7f-e14486935382-kube-api-access-hvwh8\") pod \"1776efd4-52c9-4579-aa7f-e14486935382\" (UID: \"1776efd4-52c9-4579-aa7f-e14486935382\") " Dec 10 19:53:13 crc kubenswrapper[4828]: I1210 19:53:13.988467 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1776efd4-52c9-4579-aa7f-e14486935382-utilities" (OuterVolumeSpecName: "utilities") pod "1776efd4-52c9-4579-aa7f-e14486935382" (UID: "1776efd4-52c9-4579-aa7f-e14486935382"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:53:13 crc kubenswrapper[4828]: I1210 19:53:13.995752 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1776efd4-52c9-4579-aa7f-e14486935382-kube-api-access-hvwh8" (OuterVolumeSpecName: "kube-api-access-hvwh8") pod "1776efd4-52c9-4579-aa7f-e14486935382" (UID: "1776efd4-52c9-4579-aa7f-e14486935382"). InnerVolumeSpecName "kube-api-access-hvwh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.090925 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1776efd4-52c9-4579-aa7f-e14486935382-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.090963 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvwh8\" (UniqueName: \"kubernetes.io/projected/1776efd4-52c9-4579-aa7f-e14486935382-kube-api-access-hvwh8\") on node \"crc\" DevicePath \"\"" Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.127107 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1776efd4-52c9-4579-aa7f-e14486935382-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1776efd4-52c9-4579-aa7f-e14486935382" (UID: "1776efd4-52c9-4579-aa7f-e14486935382"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.193719 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1776efd4-52c9-4579-aa7f-e14486935382-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.270478 4828 generic.go:334] "Generic (PLEG): container finished" podID="1776efd4-52c9-4579-aa7f-e14486935382" containerID="6da6e2fec5c4ddce8e42f1e7c252f4b9fc3fb11ea86a96b17da67780118fd51a" exitCode=0 Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.270569 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8k7z5" event={"ID":"1776efd4-52c9-4579-aa7f-e14486935382","Type":"ContainerDied","Data":"6da6e2fec5c4ddce8e42f1e7c252f4b9fc3fb11ea86a96b17da67780118fd51a"} Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.270615 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8k7z5" event={"ID":"1776efd4-52c9-4579-aa7f-e14486935382","Type":"ContainerDied","Data":"5ca375be6afc7d0f7e73db00fdb3f09b3ed7a5f26f9fcb3f4d80880b91defef6"} Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.270645 4828 scope.go:117] "RemoveContainer" containerID="6da6e2fec5c4ddce8e42f1e7c252f4b9fc3fb11ea86a96b17da67780118fd51a" Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.271025 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8k7z5" Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.323783 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8k7z5"] Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.324849 4828 scope.go:117] "RemoveContainer" containerID="af2c903e8d75ef0dc2c3ff6bb45ed1a1802ded842511513832e24a91fddd0696" Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.336333 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8k7z5"] Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.359519 4828 scope.go:117] "RemoveContainer" containerID="763c9b3ddb8ad3ea3ab1bf33ea297c216c60a5a581cd587139febbd215b4df82" Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.408966 4828 scope.go:117] "RemoveContainer" containerID="6da6e2fec5c4ddce8e42f1e7c252f4b9fc3fb11ea86a96b17da67780118fd51a" Dec 10 19:53:14 crc kubenswrapper[4828]: E1210 19:53:14.409505 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6da6e2fec5c4ddce8e42f1e7c252f4b9fc3fb11ea86a96b17da67780118fd51a\": container with ID starting with 6da6e2fec5c4ddce8e42f1e7c252f4b9fc3fb11ea86a96b17da67780118fd51a not found: ID does not exist" containerID="6da6e2fec5c4ddce8e42f1e7c252f4b9fc3fb11ea86a96b17da67780118fd51a" Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.409536 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6da6e2fec5c4ddce8e42f1e7c252f4b9fc3fb11ea86a96b17da67780118fd51a"} err="failed to get container status \"6da6e2fec5c4ddce8e42f1e7c252f4b9fc3fb11ea86a96b17da67780118fd51a\": rpc error: code = NotFound desc = could not find container \"6da6e2fec5c4ddce8e42f1e7c252f4b9fc3fb11ea86a96b17da67780118fd51a\": container with ID starting with 6da6e2fec5c4ddce8e42f1e7c252f4b9fc3fb11ea86a96b17da67780118fd51a not found: ID does not exist" Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.409561 4828 scope.go:117] "RemoveContainer" containerID="af2c903e8d75ef0dc2c3ff6bb45ed1a1802ded842511513832e24a91fddd0696" Dec 10 19:53:14 crc kubenswrapper[4828]: E1210 19:53:14.410081 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af2c903e8d75ef0dc2c3ff6bb45ed1a1802ded842511513832e24a91fddd0696\": container with ID starting with af2c903e8d75ef0dc2c3ff6bb45ed1a1802ded842511513832e24a91fddd0696 not found: ID does not exist" containerID="af2c903e8d75ef0dc2c3ff6bb45ed1a1802ded842511513832e24a91fddd0696" Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.410111 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af2c903e8d75ef0dc2c3ff6bb45ed1a1802ded842511513832e24a91fddd0696"} err="failed to get container status \"af2c903e8d75ef0dc2c3ff6bb45ed1a1802ded842511513832e24a91fddd0696\": rpc error: code = NotFound desc = could not find container \"af2c903e8d75ef0dc2c3ff6bb45ed1a1802ded842511513832e24a91fddd0696\": container with ID starting with af2c903e8d75ef0dc2c3ff6bb45ed1a1802ded842511513832e24a91fddd0696 not found: ID does not exist" Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.410126 4828 scope.go:117] "RemoveContainer" containerID="763c9b3ddb8ad3ea3ab1bf33ea297c216c60a5a581cd587139febbd215b4df82" Dec 10 19:53:14 crc kubenswrapper[4828]: E1210 19:53:14.410380 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"763c9b3ddb8ad3ea3ab1bf33ea297c216c60a5a581cd587139febbd215b4df82\": container with ID starting with 763c9b3ddb8ad3ea3ab1bf33ea297c216c60a5a581cd587139febbd215b4df82 not found: ID does not exist" containerID="763c9b3ddb8ad3ea3ab1bf33ea297c216c60a5a581cd587139febbd215b4df82" Dec 10 19:53:14 crc kubenswrapper[4828]: I1210 19:53:14.410487 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"763c9b3ddb8ad3ea3ab1bf33ea297c216c60a5a581cd587139febbd215b4df82"} err="failed to get container status \"763c9b3ddb8ad3ea3ab1bf33ea297c216c60a5a581cd587139febbd215b4df82\": rpc error: code = NotFound desc = could not find container \"763c9b3ddb8ad3ea3ab1bf33ea297c216c60a5a581cd587139febbd215b4df82\": container with ID starting with 763c9b3ddb8ad3ea3ab1bf33ea297c216c60a5a581cd587139febbd215b4df82 not found: ID does not exist" Dec 10 19:53:15 crc kubenswrapper[4828]: I1210 19:53:15.802528 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1776efd4-52c9-4579-aa7f-e14486935382" path="/var/lib/kubelet/pods/1776efd4-52c9-4579-aa7f-e14486935382/volumes" Dec 10 19:53:56 crc kubenswrapper[4828]: I1210 19:53:56.382308 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t2j5j"] Dec 10 19:53:56 crc kubenswrapper[4828]: E1210 19:53:56.384352 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1776efd4-52c9-4579-aa7f-e14486935382" containerName="registry-server" Dec 10 19:53:56 crc kubenswrapper[4828]: I1210 19:53:56.384388 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1776efd4-52c9-4579-aa7f-e14486935382" containerName="registry-server" Dec 10 19:53:56 crc kubenswrapper[4828]: E1210 19:53:56.384415 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1776efd4-52c9-4579-aa7f-e14486935382" containerName="extract-utilities" Dec 10 19:53:56 crc kubenswrapper[4828]: I1210 19:53:56.384424 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1776efd4-52c9-4579-aa7f-e14486935382" containerName="extract-utilities" Dec 10 19:53:56 crc kubenswrapper[4828]: E1210 19:53:56.384454 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1776efd4-52c9-4579-aa7f-e14486935382" containerName="extract-content" Dec 10 19:53:56 crc kubenswrapper[4828]: I1210 19:53:56.384462 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="1776efd4-52c9-4579-aa7f-e14486935382" containerName="extract-content" Dec 10 19:53:56 crc kubenswrapper[4828]: I1210 19:53:56.384825 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="1776efd4-52c9-4579-aa7f-e14486935382" containerName="registry-server" Dec 10 19:53:56 crc kubenswrapper[4828]: I1210 19:53:56.386997 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2j5j" Dec 10 19:53:56 crc kubenswrapper[4828]: I1210 19:53:56.400816 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t2j5j"] Dec 10 19:53:56 crc kubenswrapper[4828]: I1210 19:53:56.487290 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf5d52c3-b823-4231-b629-57ee20b609e4-utilities\") pod \"certified-operators-t2j5j\" (UID: \"bf5d52c3-b823-4231-b629-57ee20b609e4\") " pod="openshift-marketplace/certified-operators-t2j5j" Dec 10 19:53:56 crc kubenswrapper[4828]: I1210 19:53:56.487718 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j6gf\" (UniqueName: \"kubernetes.io/projected/bf5d52c3-b823-4231-b629-57ee20b609e4-kube-api-access-6j6gf\") pod \"certified-operators-t2j5j\" (UID: \"bf5d52c3-b823-4231-b629-57ee20b609e4\") " pod="openshift-marketplace/certified-operators-t2j5j" Dec 10 19:53:56 crc kubenswrapper[4828]: I1210 19:53:56.487747 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf5d52c3-b823-4231-b629-57ee20b609e4-catalog-content\") pod \"certified-operators-t2j5j\" (UID: \"bf5d52c3-b823-4231-b629-57ee20b609e4\") " pod="openshift-marketplace/certified-operators-t2j5j" Dec 10 19:53:56 crc kubenswrapper[4828]: I1210 19:53:56.590036 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf5d52c3-b823-4231-b629-57ee20b609e4-utilities\") pod \"certified-operators-t2j5j\" (UID: \"bf5d52c3-b823-4231-b629-57ee20b609e4\") " pod="openshift-marketplace/certified-operators-t2j5j" Dec 10 19:53:56 crc kubenswrapper[4828]: I1210 19:53:56.590140 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j6gf\" (UniqueName: \"kubernetes.io/projected/bf5d52c3-b823-4231-b629-57ee20b609e4-kube-api-access-6j6gf\") pod \"certified-operators-t2j5j\" (UID: \"bf5d52c3-b823-4231-b629-57ee20b609e4\") " pod="openshift-marketplace/certified-operators-t2j5j" Dec 10 19:53:56 crc kubenswrapper[4828]: I1210 19:53:56.590161 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf5d52c3-b823-4231-b629-57ee20b609e4-catalog-content\") pod \"certified-operators-t2j5j\" (UID: \"bf5d52c3-b823-4231-b629-57ee20b609e4\") " pod="openshift-marketplace/certified-operators-t2j5j" Dec 10 19:53:56 crc kubenswrapper[4828]: I1210 19:53:56.590747 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf5d52c3-b823-4231-b629-57ee20b609e4-utilities\") pod \"certified-operators-t2j5j\" (UID: \"bf5d52c3-b823-4231-b629-57ee20b609e4\") " pod="openshift-marketplace/certified-operators-t2j5j" Dec 10 19:53:56 crc kubenswrapper[4828]: I1210 19:53:56.590758 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf5d52c3-b823-4231-b629-57ee20b609e4-catalog-content\") pod \"certified-operators-t2j5j\" (UID: \"bf5d52c3-b823-4231-b629-57ee20b609e4\") " pod="openshift-marketplace/certified-operators-t2j5j" Dec 10 19:53:56 crc kubenswrapper[4828]: I1210 19:53:56.610787 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j6gf\" (UniqueName: \"kubernetes.io/projected/bf5d52c3-b823-4231-b629-57ee20b609e4-kube-api-access-6j6gf\") pod \"certified-operators-t2j5j\" (UID: \"bf5d52c3-b823-4231-b629-57ee20b609e4\") " pod="openshift-marketplace/certified-operators-t2j5j" Dec 10 19:53:56 crc kubenswrapper[4828]: I1210 19:53:56.720409 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2j5j" Dec 10 19:53:57 crc kubenswrapper[4828]: I1210 19:53:57.252959 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t2j5j"] Dec 10 19:53:57 crc kubenswrapper[4828]: I1210 19:53:57.864496 4828 generic.go:334] "Generic (PLEG): container finished" podID="bf5d52c3-b823-4231-b629-57ee20b609e4" containerID="405ea795dc0ba0929a32b1e5dfa53e34ad23b00a09d1ddcfa2edfc4d21d5cf6f" exitCode=0 Dec 10 19:53:57 crc kubenswrapper[4828]: I1210 19:53:57.865203 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2j5j" event={"ID":"bf5d52c3-b823-4231-b629-57ee20b609e4","Type":"ContainerDied","Data":"405ea795dc0ba0929a32b1e5dfa53e34ad23b00a09d1ddcfa2edfc4d21d5cf6f"} Dec 10 19:53:57 crc kubenswrapper[4828]: I1210 19:53:57.865238 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2j5j" event={"ID":"bf5d52c3-b823-4231-b629-57ee20b609e4","Type":"ContainerStarted","Data":"62b66c8f2546f6311f13a2d461803f9dcb12f9c9f6d6bcdae285bc90f51c50dd"} Dec 10 19:53:58 crc kubenswrapper[4828]: I1210 19:53:58.875551 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2j5j" event={"ID":"bf5d52c3-b823-4231-b629-57ee20b609e4","Type":"ContainerStarted","Data":"72df7dc45dafa4cedf9387e408d51c6cc4ee26735681ea1dbbe67067f624186c"} Dec 10 19:53:59 crc kubenswrapper[4828]: I1210 19:53:59.889628 4828 generic.go:334] "Generic (PLEG): container finished" podID="bf5d52c3-b823-4231-b629-57ee20b609e4" containerID="72df7dc45dafa4cedf9387e408d51c6cc4ee26735681ea1dbbe67067f624186c" exitCode=0 Dec 10 19:53:59 crc kubenswrapper[4828]: I1210 19:53:59.889679 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2j5j" event={"ID":"bf5d52c3-b823-4231-b629-57ee20b609e4","Type":"ContainerDied","Data":"72df7dc45dafa4cedf9387e408d51c6cc4ee26735681ea1dbbe67067f624186c"} Dec 10 19:54:00 crc kubenswrapper[4828]: I1210 19:54:00.900019 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2j5j" event={"ID":"bf5d52c3-b823-4231-b629-57ee20b609e4","Type":"ContainerStarted","Data":"8a8b351a97e2b900d138f4056986f8175712b86f7ced5c48e5f958f835f3ccf6"} Dec 10 19:54:00 crc kubenswrapper[4828]: I1210 19:54:00.929562 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t2j5j" podStartSLOduration=2.491865884 podStartE2EDuration="4.929538185s" podCreationTimestamp="2025-12-10 19:53:56 +0000 UTC" firstStartedPulling="2025-12-10 19:53:57.867521012 +0000 UTC m=+3518.378132007" lastFinishedPulling="2025-12-10 19:54:00.305193263 +0000 UTC m=+3520.815804308" observedRunningTime="2025-12-10 19:54:00.920233376 +0000 UTC m=+3521.430844381" watchObservedRunningTime="2025-12-10 19:54:00.929538185 +0000 UTC m=+3521.440149200" Dec 10 19:54:06 crc kubenswrapper[4828]: I1210 19:54:06.721670 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t2j5j" Dec 10 19:54:06 crc kubenswrapper[4828]: I1210 19:54:06.722257 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t2j5j" Dec 10 19:54:06 crc kubenswrapper[4828]: I1210 19:54:06.782357 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t2j5j" Dec 10 19:54:07 crc kubenswrapper[4828]: I1210 19:54:07.120481 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t2j5j" Dec 10 19:54:07 crc kubenswrapper[4828]: I1210 19:54:07.201360 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t2j5j"] Dec 10 19:54:08 crc kubenswrapper[4828]: I1210 19:54:08.995580 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t2j5j" podUID="bf5d52c3-b823-4231-b629-57ee20b609e4" containerName="registry-server" containerID="cri-o://8a8b351a97e2b900d138f4056986f8175712b86f7ced5c48e5f958f835f3ccf6" gracePeriod=2 Dec 10 19:54:10 crc kubenswrapper[4828]: I1210 19:54:10.011982 4828 generic.go:334] "Generic (PLEG): container finished" podID="bf5d52c3-b823-4231-b629-57ee20b609e4" containerID="8a8b351a97e2b900d138f4056986f8175712b86f7ced5c48e5f958f835f3ccf6" exitCode=0 Dec 10 19:54:10 crc kubenswrapper[4828]: I1210 19:54:10.012024 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2j5j" event={"ID":"bf5d52c3-b823-4231-b629-57ee20b609e4","Type":"ContainerDied","Data":"8a8b351a97e2b900d138f4056986f8175712b86f7ced5c48e5f958f835f3ccf6"} Dec 10 19:54:10 crc kubenswrapper[4828]: I1210 19:54:10.664570 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2j5j" Dec 10 19:54:10 crc kubenswrapper[4828]: I1210 19:54:10.714075 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf5d52c3-b823-4231-b629-57ee20b609e4-catalog-content\") pod \"bf5d52c3-b823-4231-b629-57ee20b609e4\" (UID: \"bf5d52c3-b823-4231-b629-57ee20b609e4\") " Dec 10 19:54:10 crc kubenswrapper[4828]: I1210 19:54:10.714211 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf5d52c3-b823-4231-b629-57ee20b609e4-utilities\") pod \"bf5d52c3-b823-4231-b629-57ee20b609e4\" (UID: \"bf5d52c3-b823-4231-b629-57ee20b609e4\") " Dec 10 19:54:10 crc kubenswrapper[4828]: I1210 19:54:10.714471 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6j6gf\" (UniqueName: \"kubernetes.io/projected/bf5d52c3-b823-4231-b629-57ee20b609e4-kube-api-access-6j6gf\") pod \"bf5d52c3-b823-4231-b629-57ee20b609e4\" (UID: \"bf5d52c3-b823-4231-b629-57ee20b609e4\") " Dec 10 19:54:10 crc kubenswrapper[4828]: I1210 19:54:10.715771 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf5d52c3-b823-4231-b629-57ee20b609e4-utilities" (OuterVolumeSpecName: "utilities") pod "bf5d52c3-b823-4231-b629-57ee20b609e4" (UID: "bf5d52c3-b823-4231-b629-57ee20b609e4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:54:10 crc kubenswrapper[4828]: I1210 19:54:10.721014 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf5d52c3-b823-4231-b629-57ee20b609e4-kube-api-access-6j6gf" (OuterVolumeSpecName: "kube-api-access-6j6gf") pod "bf5d52c3-b823-4231-b629-57ee20b609e4" (UID: "bf5d52c3-b823-4231-b629-57ee20b609e4"). InnerVolumeSpecName "kube-api-access-6j6gf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:54:10 crc kubenswrapper[4828]: I1210 19:54:10.767683 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf5d52c3-b823-4231-b629-57ee20b609e4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf5d52c3-b823-4231-b629-57ee20b609e4" (UID: "bf5d52c3-b823-4231-b629-57ee20b609e4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:54:10 crc kubenswrapper[4828]: I1210 19:54:10.817155 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6j6gf\" (UniqueName: \"kubernetes.io/projected/bf5d52c3-b823-4231-b629-57ee20b609e4-kube-api-access-6j6gf\") on node \"crc\" DevicePath \"\"" Dec 10 19:54:10 crc kubenswrapper[4828]: I1210 19:54:10.817191 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf5d52c3-b823-4231-b629-57ee20b609e4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:54:10 crc kubenswrapper[4828]: I1210 19:54:10.817202 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf5d52c3-b823-4231-b629-57ee20b609e4-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:54:11 crc kubenswrapper[4828]: I1210 19:54:11.026118 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t2j5j" event={"ID":"bf5d52c3-b823-4231-b629-57ee20b609e4","Type":"ContainerDied","Data":"62b66c8f2546f6311f13a2d461803f9dcb12f9c9f6d6bcdae285bc90f51c50dd"} Dec 10 19:54:11 crc kubenswrapper[4828]: I1210 19:54:11.026172 4828 scope.go:117] "RemoveContainer" containerID="8a8b351a97e2b900d138f4056986f8175712b86f7ced5c48e5f958f835f3ccf6" Dec 10 19:54:11 crc kubenswrapper[4828]: I1210 19:54:11.026215 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t2j5j" Dec 10 19:54:11 crc kubenswrapper[4828]: I1210 19:54:11.061147 4828 scope.go:117] "RemoveContainer" containerID="72df7dc45dafa4cedf9387e408d51c6cc4ee26735681ea1dbbe67067f624186c" Dec 10 19:54:11 crc kubenswrapper[4828]: I1210 19:54:11.068712 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t2j5j"] Dec 10 19:54:11 crc kubenswrapper[4828]: I1210 19:54:11.081713 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t2j5j"] Dec 10 19:54:11 crc kubenswrapper[4828]: I1210 19:54:11.089272 4828 scope.go:117] "RemoveContainer" containerID="405ea795dc0ba0929a32b1e5dfa53e34ad23b00a09d1ddcfa2edfc4d21d5cf6f" Dec 10 19:54:11 crc kubenswrapper[4828]: I1210 19:54:11.808298 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf5d52c3-b823-4231-b629-57ee20b609e4" path="/var/lib/kubelet/pods/bf5d52c3-b823-4231-b629-57ee20b609e4/volumes" Dec 10 19:54:21 crc kubenswrapper[4828]: I1210 19:54:21.230435 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:54:21 crc kubenswrapper[4828]: I1210 19:54:21.231032 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:54:51 crc kubenswrapper[4828]: I1210 19:54:51.231137 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:54:51 crc kubenswrapper[4828]: I1210 19:54:51.231692 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:55:21 crc kubenswrapper[4828]: I1210 19:55:21.230162 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 19:55:21 crc kubenswrapper[4828]: I1210 19:55:21.230669 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 19:55:21 crc kubenswrapper[4828]: I1210 19:55:21.230718 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 19:55:21 crc kubenswrapper[4828]: I1210 19:55:21.231633 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 19:55:21 crc kubenswrapper[4828]: I1210 19:55:21.231685 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" gracePeriod=600 Dec 10 19:55:21 crc kubenswrapper[4828]: I1210 19:55:21.786068 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" exitCode=0 Dec 10 19:55:21 crc kubenswrapper[4828]: I1210 19:55:21.786404 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd"} Dec 10 19:55:21 crc kubenswrapper[4828]: I1210 19:55:21.786446 4828 scope.go:117] "RemoveContainer" containerID="558207227d2901f2b3b87b61b1e8400a08f602aa55d126b771706aa1bef786c1" Dec 10 19:55:21 crc kubenswrapper[4828]: E1210 19:55:21.894731 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:55:22 crc kubenswrapper[4828]: I1210 19:55:22.801200 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:55:22 crc kubenswrapper[4828]: E1210 19:55:22.801980 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:55:34 crc kubenswrapper[4828]: I1210 19:55:34.789544 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:55:34 crc kubenswrapper[4828]: E1210 19:55:34.790375 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:55:46 crc kubenswrapper[4828]: I1210 19:55:46.322682 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vjb6f"] Dec 10 19:55:46 crc kubenswrapper[4828]: E1210 19:55:46.323723 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf5d52c3-b823-4231-b629-57ee20b609e4" containerName="extract-utilities" Dec 10 19:55:46 crc kubenswrapper[4828]: I1210 19:55:46.323738 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf5d52c3-b823-4231-b629-57ee20b609e4" containerName="extract-utilities" Dec 10 19:55:46 crc kubenswrapper[4828]: E1210 19:55:46.323768 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf5d52c3-b823-4231-b629-57ee20b609e4" containerName="registry-server" Dec 10 19:55:46 crc kubenswrapper[4828]: I1210 19:55:46.323774 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf5d52c3-b823-4231-b629-57ee20b609e4" containerName="registry-server" Dec 10 19:55:46 crc kubenswrapper[4828]: E1210 19:55:46.323783 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf5d52c3-b823-4231-b629-57ee20b609e4" containerName="extract-content" Dec 10 19:55:46 crc kubenswrapper[4828]: I1210 19:55:46.323789 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf5d52c3-b823-4231-b629-57ee20b609e4" containerName="extract-content" Dec 10 19:55:46 crc kubenswrapper[4828]: I1210 19:55:46.324024 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf5d52c3-b823-4231-b629-57ee20b609e4" containerName="registry-server" Dec 10 19:55:46 crc kubenswrapper[4828]: I1210 19:55:46.330057 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vjb6f" Dec 10 19:55:46 crc kubenswrapper[4828]: I1210 19:55:46.364895 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vjb6f"] Dec 10 19:55:46 crc kubenswrapper[4828]: I1210 19:55:46.418852 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ced0629-eb0b-4acd-b2e7-d521f29536f9-utilities\") pod \"community-operators-vjb6f\" (UID: \"5ced0629-eb0b-4acd-b2e7-d521f29536f9\") " pod="openshift-marketplace/community-operators-vjb6f" Dec 10 19:55:46 crc kubenswrapper[4828]: I1210 19:55:46.419262 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ced0629-eb0b-4acd-b2e7-d521f29536f9-catalog-content\") pod \"community-operators-vjb6f\" (UID: \"5ced0629-eb0b-4acd-b2e7-d521f29536f9\") " pod="openshift-marketplace/community-operators-vjb6f" Dec 10 19:55:46 crc kubenswrapper[4828]: I1210 19:55:46.419306 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdjd9\" (UniqueName: \"kubernetes.io/projected/5ced0629-eb0b-4acd-b2e7-d521f29536f9-kube-api-access-wdjd9\") pod \"community-operators-vjb6f\" (UID: \"5ced0629-eb0b-4acd-b2e7-d521f29536f9\") " pod="openshift-marketplace/community-operators-vjb6f" Dec 10 19:55:46 crc kubenswrapper[4828]: I1210 19:55:46.521170 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ced0629-eb0b-4acd-b2e7-d521f29536f9-utilities\") pod \"community-operators-vjb6f\" (UID: \"5ced0629-eb0b-4acd-b2e7-d521f29536f9\") " pod="openshift-marketplace/community-operators-vjb6f" Dec 10 19:55:46 crc kubenswrapper[4828]: I1210 19:55:46.521502 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ced0629-eb0b-4acd-b2e7-d521f29536f9-catalog-content\") pod \"community-operators-vjb6f\" (UID: \"5ced0629-eb0b-4acd-b2e7-d521f29536f9\") " pod="openshift-marketplace/community-operators-vjb6f" Dec 10 19:55:46 crc kubenswrapper[4828]: I1210 19:55:46.521608 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdjd9\" (UniqueName: \"kubernetes.io/projected/5ced0629-eb0b-4acd-b2e7-d521f29536f9-kube-api-access-wdjd9\") pod \"community-operators-vjb6f\" (UID: \"5ced0629-eb0b-4acd-b2e7-d521f29536f9\") " pod="openshift-marketplace/community-operators-vjb6f" Dec 10 19:55:46 crc kubenswrapper[4828]: I1210 19:55:46.521759 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ced0629-eb0b-4acd-b2e7-d521f29536f9-utilities\") pod \"community-operators-vjb6f\" (UID: \"5ced0629-eb0b-4acd-b2e7-d521f29536f9\") " pod="openshift-marketplace/community-operators-vjb6f" Dec 10 19:55:46 crc kubenswrapper[4828]: I1210 19:55:46.521955 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ced0629-eb0b-4acd-b2e7-d521f29536f9-catalog-content\") pod \"community-operators-vjb6f\" (UID: \"5ced0629-eb0b-4acd-b2e7-d521f29536f9\") " pod="openshift-marketplace/community-operators-vjb6f" Dec 10 19:55:46 crc kubenswrapper[4828]: I1210 19:55:46.545871 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdjd9\" (UniqueName: \"kubernetes.io/projected/5ced0629-eb0b-4acd-b2e7-d521f29536f9-kube-api-access-wdjd9\") pod \"community-operators-vjb6f\" (UID: \"5ced0629-eb0b-4acd-b2e7-d521f29536f9\") " pod="openshift-marketplace/community-operators-vjb6f" Dec 10 19:55:46 crc kubenswrapper[4828]: I1210 19:55:46.662185 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vjb6f" Dec 10 19:55:47 crc kubenswrapper[4828]: I1210 19:55:47.272929 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vjb6f"] Dec 10 19:55:47 crc kubenswrapper[4828]: I1210 19:55:47.790509 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:55:47 crc kubenswrapper[4828]: E1210 19:55:47.791012 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:55:48 crc kubenswrapper[4828]: I1210 19:55:48.078222 4828 generic.go:334] "Generic (PLEG): container finished" podID="5ced0629-eb0b-4acd-b2e7-d521f29536f9" containerID="ba5e2910d9d5a46a0b7fe9a8c99b43c73841a601c80be0bebac107a17783f181" exitCode=0 Dec 10 19:55:48 crc kubenswrapper[4828]: I1210 19:55:48.078316 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjb6f" event={"ID":"5ced0629-eb0b-4acd-b2e7-d521f29536f9","Type":"ContainerDied","Data":"ba5e2910d9d5a46a0b7fe9a8c99b43c73841a601c80be0bebac107a17783f181"} Dec 10 19:55:48 crc kubenswrapper[4828]: I1210 19:55:48.078496 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjb6f" event={"ID":"5ced0629-eb0b-4acd-b2e7-d521f29536f9","Type":"ContainerStarted","Data":"a8e6e6651919ad77895d884df5e965cdefffd8ce4f33dda4048320354b1d6a1d"} Dec 10 19:55:50 crc kubenswrapper[4828]: I1210 19:55:50.103240 4828 generic.go:334] "Generic (PLEG): container finished" podID="5ced0629-eb0b-4acd-b2e7-d521f29536f9" containerID="ee33d27c00ec077efa9d8dcda88876747e17ac1d7b5df6d251569928b8d52280" exitCode=0 Dec 10 19:55:50 crc kubenswrapper[4828]: I1210 19:55:50.103403 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjb6f" event={"ID":"5ced0629-eb0b-4acd-b2e7-d521f29536f9","Type":"ContainerDied","Data":"ee33d27c00ec077efa9d8dcda88876747e17ac1d7b5df6d251569928b8d52280"} Dec 10 19:55:51 crc kubenswrapper[4828]: I1210 19:55:51.116342 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjb6f" event={"ID":"5ced0629-eb0b-4acd-b2e7-d521f29536f9","Type":"ContainerStarted","Data":"28f5f46e070331407518d57b5e8b430a56fbb2c184d6d881ebf959a24d665287"} Dec 10 19:55:51 crc kubenswrapper[4828]: I1210 19:55:51.139860 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vjb6f" podStartSLOduration=2.545178027 podStartE2EDuration="5.139834973s" podCreationTimestamp="2025-12-10 19:55:46 +0000 UTC" firstStartedPulling="2025-12-10 19:55:48.082485655 +0000 UTC m=+3628.593096660" lastFinishedPulling="2025-12-10 19:55:50.677142601 +0000 UTC m=+3631.187753606" observedRunningTime="2025-12-10 19:55:51.133949736 +0000 UTC m=+3631.644560741" watchObservedRunningTime="2025-12-10 19:55:51.139834973 +0000 UTC m=+3631.650445978" Dec 10 19:55:55 crc kubenswrapper[4828]: I1210 19:55:55.492573 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-726nd"] Dec 10 19:55:55 crc kubenswrapper[4828]: I1210 19:55:55.497592 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-726nd" Dec 10 19:55:55 crc kubenswrapper[4828]: I1210 19:55:55.512318 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-726nd"] Dec 10 19:55:55 crc kubenswrapper[4828]: I1210 19:55:55.556599 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2814683-d1f3-4307-a95c-94eefe173bf8-catalog-content\") pod \"redhat-marketplace-726nd\" (UID: \"b2814683-d1f3-4307-a95c-94eefe173bf8\") " pod="openshift-marketplace/redhat-marketplace-726nd" Dec 10 19:55:55 crc kubenswrapper[4828]: I1210 19:55:55.556716 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv57w\" (UniqueName: \"kubernetes.io/projected/b2814683-d1f3-4307-a95c-94eefe173bf8-kube-api-access-hv57w\") pod \"redhat-marketplace-726nd\" (UID: \"b2814683-d1f3-4307-a95c-94eefe173bf8\") " pod="openshift-marketplace/redhat-marketplace-726nd" Dec 10 19:55:55 crc kubenswrapper[4828]: I1210 19:55:55.559214 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2814683-d1f3-4307-a95c-94eefe173bf8-utilities\") pod \"redhat-marketplace-726nd\" (UID: \"b2814683-d1f3-4307-a95c-94eefe173bf8\") " pod="openshift-marketplace/redhat-marketplace-726nd" Dec 10 19:55:55 crc kubenswrapper[4828]: I1210 19:55:55.661456 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv57w\" (UniqueName: \"kubernetes.io/projected/b2814683-d1f3-4307-a95c-94eefe173bf8-kube-api-access-hv57w\") pod \"redhat-marketplace-726nd\" (UID: \"b2814683-d1f3-4307-a95c-94eefe173bf8\") " pod="openshift-marketplace/redhat-marketplace-726nd" Dec 10 19:55:55 crc kubenswrapper[4828]: I1210 19:55:55.661688 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2814683-d1f3-4307-a95c-94eefe173bf8-utilities\") pod \"redhat-marketplace-726nd\" (UID: \"b2814683-d1f3-4307-a95c-94eefe173bf8\") " pod="openshift-marketplace/redhat-marketplace-726nd" Dec 10 19:55:55 crc kubenswrapper[4828]: I1210 19:55:55.661736 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2814683-d1f3-4307-a95c-94eefe173bf8-catalog-content\") pod \"redhat-marketplace-726nd\" (UID: \"b2814683-d1f3-4307-a95c-94eefe173bf8\") " pod="openshift-marketplace/redhat-marketplace-726nd" Dec 10 19:55:55 crc kubenswrapper[4828]: I1210 19:55:55.662286 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2814683-d1f3-4307-a95c-94eefe173bf8-catalog-content\") pod \"redhat-marketplace-726nd\" (UID: \"b2814683-d1f3-4307-a95c-94eefe173bf8\") " pod="openshift-marketplace/redhat-marketplace-726nd" Dec 10 19:55:55 crc kubenswrapper[4828]: I1210 19:55:55.662570 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2814683-d1f3-4307-a95c-94eefe173bf8-utilities\") pod \"redhat-marketplace-726nd\" (UID: \"b2814683-d1f3-4307-a95c-94eefe173bf8\") " pod="openshift-marketplace/redhat-marketplace-726nd" Dec 10 19:55:55 crc kubenswrapper[4828]: I1210 19:55:55.683976 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv57w\" (UniqueName: \"kubernetes.io/projected/b2814683-d1f3-4307-a95c-94eefe173bf8-kube-api-access-hv57w\") pod \"redhat-marketplace-726nd\" (UID: \"b2814683-d1f3-4307-a95c-94eefe173bf8\") " pod="openshift-marketplace/redhat-marketplace-726nd" Dec 10 19:55:55 crc kubenswrapper[4828]: I1210 19:55:55.822258 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-726nd" Dec 10 19:55:56 crc kubenswrapper[4828]: I1210 19:55:56.373020 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-726nd"] Dec 10 19:55:56 crc kubenswrapper[4828]: I1210 19:55:56.662532 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vjb6f" Dec 10 19:55:56 crc kubenswrapper[4828]: I1210 19:55:56.662590 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vjb6f" Dec 10 19:55:56 crc kubenswrapper[4828]: I1210 19:55:56.720298 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vjb6f" Dec 10 19:55:57 crc kubenswrapper[4828]: I1210 19:55:57.215274 4828 generic.go:334] "Generic (PLEG): container finished" podID="b2814683-d1f3-4307-a95c-94eefe173bf8" containerID="cb293c341c98f7bb0ea144a5422c436c2724fbbe931cb1686b1ec7e4306513eb" exitCode=0 Dec 10 19:55:57 crc kubenswrapper[4828]: I1210 19:55:57.215324 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-726nd" event={"ID":"b2814683-d1f3-4307-a95c-94eefe173bf8","Type":"ContainerDied","Data":"cb293c341c98f7bb0ea144a5422c436c2724fbbe931cb1686b1ec7e4306513eb"} Dec 10 19:55:57 crc kubenswrapper[4828]: I1210 19:55:57.215393 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-726nd" event={"ID":"b2814683-d1f3-4307-a95c-94eefe173bf8","Type":"ContainerStarted","Data":"a9dc414bfe3602ccfeecfd5c92e13b224a3e9aa48e9a5701e425715dadfca70c"} Dec 10 19:55:57 crc kubenswrapper[4828]: I1210 19:55:57.270836 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vjb6f" Dec 10 19:55:58 crc kubenswrapper[4828]: I1210 19:55:58.231712 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-726nd" event={"ID":"b2814683-d1f3-4307-a95c-94eefe173bf8","Type":"ContainerStarted","Data":"ac423d09dad304a98658b773e736b0d212a9fc7c61136c06a18b5e3b1e83d49e"} Dec 10 19:55:59 crc kubenswrapper[4828]: I1210 19:55:59.061059 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vjb6f"] Dec 10 19:55:59 crc kubenswrapper[4828]: I1210 19:55:59.243728 4828 generic.go:334] "Generic (PLEG): container finished" podID="b2814683-d1f3-4307-a95c-94eefe173bf8" containerID="ac423d09dad304a98658b773e736b0d212a9fc7c61136c06a18b5e3b1e83d49e" exitCode=0 Dec 10 19:55:59 crc kubenswrapper[4828]: I1210 19:55:59.243873 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-726nd" event={"ID":"b2814683-d1f3-4307-a95c-94eefe173bf8","Type":"ContainerDied","Data":"ac423d09dad304a98658b773e736b0d212a9fc7c61136c06a18b5e3b1e83d49e"} Dec 10 19:55:59 crc kubenswrapper[4828]: I1210 19:55:59.244541 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vjb6f" podUID="5ced0629-eb0b-4acd-b2e7-d521f29536f9" containerName="registry-server" containerID="cri-o://28f5f46e070331407518d57b5e8b430a56fbb2c184d6d881ebf959a24d665287" gracePeriod=2 Dec 10 19:55:59 crc kubenswrapper[4828]: I1210 19:55:59.786733 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vjb6f" Dec 10 19:55:59 crc kubenswrapper[4828]: I1210 19:55:59.878363 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ced0629-eb0b-4acd-b2e7-d521f29536f9-catalog-content\") pod \"5ced0629-eb0b-4acd-b2e7-d521f29536f9\" (UID: \"5ced0629-eb0b-4acd-b2e7-d521f29536f9\") " Dec 10 19:55:59 crc kubenswrapper[4828]: I1210 19:55:59.878850 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ced0629-eb0b-4acd-b2e7-d521f29536f9-utilities\") pod \"5ced0629-eb0b-4acd-b2e7-d521f29536f9\" (UID: \"5ced0629-eb0b-4acd-b2e7-d521f29536f9\") " Dec 10 19:55:59 crc kubenswrapper[4828]: I1210 19:55:59.878933 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdjd9\" (UniqueName: \"kubernetes.io/projected/5ced0629-eb0b-4acd-b2e7-d521f29536f9-kube-api-access-wdjd9\") pod \"5ced0629-eb0b-4acd-b2e7-d521f29536f9\" (UID: \"5ced0629-eb0b-4acd-b2e7-d521f29536f9\") " Dec 10 19:55:59 crc kubenswrapper[4828]: I1210 19:55:59.879716 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ced0629-eb0b-4acd-b2e7-d521f29536f9-utilities" (OuterVolumeSpecName: "utilities") pod "5ced0629-eb0b-4acd-b2e7-d521f29536f9" (UID: "5ced0629-eb0b-4acd-b2e7-d521f29536f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:55:59 crc kubenswrapper[4828]: I1210 19:55:59.885053 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ced0629-eb0b-4acd-b2e7-d521f29536f9-kube-api-access-wdjd9" (OuterVolumeSpecName: "kube-api-access-wdjd9") pod "5ced0629-eb0b-4acd-b2e7-d521f29536f9" (UID: "5ced0629-eb0b-4acd-b2e7-d521f29536f9"). InnerVolumeSpecName "kube-api-access-wdjd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:55:59 crc kubenswrapper[4828]: I1210 19:55:59.926858 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ced0629-eb0b-4acd-b2e7-d521f29536f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ced0629-eb0b-4acd-b2e7-d521f29536f9" (UID: "5ced0629-eb0b-4acd-b2e7-d521f29536f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:55:59 crc kubenswrapper[4828]: I1210 19:55:59.982149 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ced0629-eb0b-4acd-b2e7-d521f29536f9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:55:59 crc kubenswrapper[4828]: I1210 19:55:59.982200 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ced0629-eb0b-4acd-b2e7-d521f29536f9-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:55:59 crc kubenswrapper[4828]: I1210 19:55:59.982224 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdjd9\" (UniqueName: \"kubernetes.io/projected/5ced0629-eb0b-4acd-b2e7-d521f29536f9-kube-api-access-wdjd9\") on node \"crc\" DevicePath \"\"" Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.258550 4828 generic.go:334] "Generic (PLEG): container finished" podID="5ced0629-eb0b-4acd-b2e7-d521f29536f9" containerID="28f5f46e070331407518d57b5e8b430a56fbb2c184d6d881ebf959a24d665287" exitCode=0 Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.258620 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjb6f" event={"ID":"5ced0629-eb0b-4acd-b2e7-d521f29536f9","Type":"ContainerDied","Data":"28f5f46e070331407518d57b5e8b430a56fbb2c184d6d881ebf959a24d665287"} Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.258668 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vjb6f" Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.258703 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vjb6f" event={"ID":"5ced0629-eb0b-4acd-b2e7-d521f29536f9","Type":"ContainerDied","Data":"a8e6e6651919ad77895d884df5e965cdefffd8ce4f33dda4048320354b1d6a1d"} Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.258734 4828 scope.go:117] "RemoveContainer" containerID="28f5f46e070331407518d57b5e8b430a56fbb2c184d6d881ebf959a24d665287" Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.265872 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-726nd" event={"ID":"b2814683-d1f3-4307-a95c-94eefe173bf8","Type":"ContainerStarted","Data":"f256fc967f4e84327fc3ac3238cd1d91ab6b0c98573730addb652c6874b7c83c"} Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.288378 4828 scope.go:117] "RemoveContainer" containerID="ee33d27c00ec077efa9d8dcda88876747e17ac1d7b5df6d251569928b8d52280" Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.298201 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-726nd" podStartSLOduration=2.558282744 podStartE2EDuration="5.298181331s" podCreationTimestamp="2025-12-10 19:55:55 +0000 UTC" firstStartedPulling="2025-12-10 19:55:57.218890605 +0000 UTC m=+3637.729501630" lastFinishedPulling="2025-12-10 19:55:59.958789212 +0000 UTC m=+3640.469400217" observedRunningTime="2025-12-10 19:56:00.285099621 +0000 UTC m=+3640.795710626" watchObservedRunningTime="2025-12-10 19:56:00.298181331 +0000 UTC m=+3640.808792336" Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.315256 4828 scope.go:117] "RemoveContainer" containerID="ba5e2910d9d5a46a0b7fe9a8c99b43c73841a601c80be0bebac107a17783f181" Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.323297 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vjb6f"] Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.339310 4828 scope.go:117] "RemoveContainer" containerID="28f5f46e070331407518d57b5e8b430a56fbb2c184d6d881ebf959a24d665287" Dec 10 19:56:00 crc kubenswrapper[4828]: E1210 19:56:00.339739 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28f5f46e070331407518d57b5e8b430a56fbb2c184d6d881ebf959a24d665287\": container with ID starting with 28f5f46e070331407518d57b5e8b430a56fbb2c184d6d881ebf959a24d665287 not found: ID does not exist" containerID="28f5f46e070331407518d57b5e8b430a56fbb2c184d6d881ebf959a24d665287" Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.339767 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28f5f46e070331407518d57b5e8b430a56fbb2c184d6d881ebf959a24d665287"} err="failed to get container status \"28f5f46e070331407518d57b5e8b430a56fbb2c184d6d881ebf959a24d665287\": rpc error: code = NotFound desc = could not find container \"28f5f46e070331407518d57b5e8b430a56fbb2c184d6d881ebf959a24d665287\": container with ID starting with 28f5f46e070331407518d57b5e8b430a56fbb2c184d6d881ebf959a24d665287 not found: ID does not exist" Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.339789 4828 scope.go:117] "RemoveContainer" containerID="ee33d27c00ec077efa9d8dcda88876747e17ac1d7b5df6d251569928b8d52280" Dec 10 19:56:00 crc kubenswrapper[4828]: E1210 19:56:00.340043 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee33d27c00ec077efa9d8dcda88876747e17ac1d7b5df6d251569928b8d52280\": container with ID starting with ee33d27c00ec077efa9d8dcda88876747e17ac1d7b5df6d251569928b8d52280 not found: ID does not exist" containerID="ee33d27c00ec077efa9d8dcda88876747e17ac1d7b5df6d251569928b8d52280" Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.340067 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee33d27c00ec077efa9d8dcda88876747e17ac1d7b5df6d251569928b8d52280"} err="failed to get container status \"ee33d27c00ec077efa9d8dcda88876747e17ac1d7b5df6d251569928b8d52280\": rpc error: code = NotFound desc = could not find container \"ee33d27c00ec077efa9d8dcda88876747e17ac1d7b5df6d251569928b8d52280\": container with ID starting with ee33d27c00ec077efa9d8dcda88876747e17ac1d7b5df6d251569928b8d52280 not found: ID does not exist" Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.340102 4828 scope.go:117] "RemoveContainer" containerID="ba5e2910d9d5a46a0b7fe9a8c99b43c73841a601c80be0bebac107a17783f181" Dec 10 19:56:00 crc kubenswrapper[4828]: E1210 19:56:00.340286 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba5e2910d9d5a46a0b7fe9a8c99b43c73841a601c80be0bebac107a17783f181\": container with ID starting with ba5e2910d9d5a46a0b7fe9a8c99b43c73841a601c80be0bebac107a17783f181 not found: ID does not exist" containerID="ba5e2910d9d5a46a0b7fe9a8c99b43c73841a601c80be0bebac107a17783f181" Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.340308 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba5e2910d9d5a46a0b7fe9a8c99b43c73841a601c80be0bebac107a17783f181"} err="failed to get container status \"ba5e2910d9d5a46a0b7fe9a8c99b43c73841a601c80be0bebac107a17783f181\": rpc error: code = NotFound desc = could not find container \"ba5e2910d9d5a46a0b7fe9a8c99b43c73841a601c80be0bebac107a17783f181\": container with ID starting with ba5e2910d9d5a46a0b7fe9a8c99b43c73841a601c80be0bebac107a17783f181 not found: ID does not exist" Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.342524 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vjb6f"] Dec 10 19:56:00 crc kubenswrapper[4828]: I1210 19:56:00.789629 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:56:00 crc kubenswrapper[4828]: E1210 19:56:00.790235 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:56:01 crc kubenswrapper[4828]: I1210 19:56:01.805599 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ced0629-eb0b-4acd-b2e7-d521f29536f9" path="/var/lib/kubelet/pods/5ced0629-eb0b-4acd-b2e7-d521f29536f9/volumes" Dec 10 19:56:05 crc kubenswrapper[4828]: I1210 19:56:05.822671 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-726nd" Dec 10 19:56:05 crc kubenswrapper[4828]: I1210 19:56:05.823389 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-726nd" Dec 10 19:56:05 crc kubenswrapper[4828]: I1210 19:56:05.882592 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-726nd" Dec 10 19:56:06 crc kubenswrapper[4828]: I1210 19:56:06.402044 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-726nd" Dec 10 19:56:06 crc kubenswrapper[4828]: I1210 19:56:06.452290 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-726nd"] Dec 10 19:56:08 crc kubenswrapper[4828]: I1210 19:56:08.365397 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-726nd" podUID="b2814683-d1f3-4307-a95c-94eefe173bf8" containerName="registry-server" containerID="cri-o://f256fc967f4e84327fc3ac3238cd1d91ab6b0c98573730addb652c6874b7c83c" gracePeriod=2 Dec 10 19:56:08 crc kubenswrapper[4828]: I1210 19:56:08.901450 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-726nd" Dec 10 19:56:08 crc kubenswrapper[4828]: I1210 19:56:08.998569 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv57w\" (UniqueName: \"kubernetes.io/projected/b2814683-d1f3-4307-a95c-94eefe173bf8-kube-api-access-hv57w\") pod \"b2814683-d1f3-4307-a95c-94eefe173bf8\" (UID: \"b2814683-d1f3-4307-a95c-94eefe173bf8\") " Dec 10 19:56:08 crc kubenswrapper[4828]: I1210 19:56:08.998958 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2814683-d1f3-4307-a95c-94eefe173bf8-utilities\") pod \"b2814683-d1f3-4307-a95c-94eefe173bf8\" (UID: \"b2814683-d1f3-4307-a95c-94eefe173bf8\") " Dec 10 19:56:08 crc kubenswrapper[4828]: I1210 19:56:08.999040 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2814683-d1f3-4307-a95c-94eefe173bf8-catalog-content\") pod \"b2814683-d1f3-4307-a95c-94eefe173bf8\" (UID: \"b2814683-d1f3-4307-a95c-94eefe173bf8\") " Dec 10 19:56:08 crc kubenswrapper[4828]: I1210 19:56:08.999775 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2814683-d1f3-4307-a95c-94eefe173bf8-utilities" (OuterVolumeSpecName: "utilities") pod "b2814683-d1f3-4307-a95c-94eefe173bf8" (UID: "b2814683-d1f3-4307-a95c-94eefe173bf8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.004416 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2814683-d1f3-4307-a95c-94eefe173bf8-kube-api-access-hv57w" (OuterVolumeSpecName: "kube-api-access-hv57w") pod "b2814683-d1f3-4307-a95c-94eefe173bf8" (UID: "b2814683-d1f3-4307-a95c-94eefe173bf8"). InnerVolumeSpecName "kube-api-access-hv57w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.021225 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2814683-d1f3-4307-a95c-94eefe173bf8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2814683-d1f3-4307-a95c-94eefe173bf8" (UID: "b2814683-d1f3-4307-a95c-94eefe173bf8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.101629 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2814683-d1f3-4307-a95c-94eefe173bf8-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.101659 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2814683-d1f3-4307-a95c-94eefe173bf8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.101676 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv57w\" (UniqueName: \"kubernetes.io/projected/b2814683-d1f3-4307-a95c-94eefe173bf8-kube-api-access-hv57w\") on node \"crc\" DevicePath \"\"" Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.385441 4828 generic.go:334] "Generic (PLEG): container finished" podID="b2814683-d1f3-4307-a95c-94eefe173bf8" containerID="f256fc967f4e84327fc3ac3238cd1d91ab6b0c98573730addb652c6874b7c83c" exitCode=0 Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.385497 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-726nd" Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.385544 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-726nd" event={"ID":"b2814683-d1f3-4307-a95c-94eefe173bf8","Type":"ContainerDied","Data":"f256fc967f4e84327fc3ac3238cd1d91ab6b0c98573730addb652c6874b7c83c"} Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.386667 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-726nd" event={"ID":"b2814683-d1f3-4307-a95c-94eefe173bf8","Type":"ContainerDied","Data":"a9dc414bfe3602ccfeecfd5c92e13b224a3e9aa48e9a5701e425715dadfca70c"} Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.386704 4828 scope.go:117] "RemoveContainer" containerID="f256fc967f4e84327fc3ac3238cd1d91ab6b0c98573730addb652c6874b7c83c" Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.431715 4828 scope.go:117] "RemoveContainer" containerID="ac423d09dad304a98658b773e736b0d212a9fc7c61136c06a18b5e3b1e83d49e" Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.445285 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-726nd"] Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.458152 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-726nd"] Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.471570 4828 scope.go:117] "RemoveContainer" containerID="cb293c341c98f7bb0ea144a5422c436c2724fbbe931cb1686b1ec7e4306513eb" Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.541250 4828 scope.go:117] "RemoveContainer" containerID="f256fc967f4e84327fc3ac3238cd1d91ab6b0c98573730addb652c6874b7c83c" Dec 10 19:56:09 crc kubenswrapper[4828]: E1210 19:56:09.541687 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f256fc967f4e84327fc3ac3238cd1d91ab6b0c98573730addb652c6874b7c83c\": container with ID starting with f256fc967f4e84327fc3ac3238cd1d91ab6b0c98573730addb652c6874b7c83c not found: ID does not exist" containerID="f256fc967f4e84327fc3ac3238cd1d91ab6b0c98573730addb652c6874b7c83c" Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.541737 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f256fc967f4e84327fc3ac3238cd1d91ab6b0c98573730addb652c6874b7c83c"} err="failed to get container status \"f256fc967f4e84327fc3ac3238cd1d91ab6b0c98573730addb652c6874b7c83c\": rpc error: code = NotFound desc = could not find container \"f256fc967f4e84327fc3ac3238cd1d91ab6b0c98573730addb652c6874b7c83c\": container with ID starting with f256fc967f4e84327fc3ac3238cd1d91ab6b0c98573730addb652c6874b7c83c not found: ID does not exist" Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.541764 4828 scope.go:117] "RemoveContainer" containerID="ac423d09dad304a98658b773e736b0d212a9fc7c61136c06a18b5e3b1e83d49e" Dec 10 19:56:09 crc kubenswrapper[4828]: E1210 19:56:09.542146 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac423d09dad304a98658b773e736b0d212a9fc7c61136c06a18b5e3b1e83d49e\": container with ID starting with ac423d09dad304a98658b773e736b0d212a9fc7c61136c06a18b5e3b1e83d49e not found: ID does not exist" containerID="ac423d09dad304a98658b773e736b0d212a9fc7c61136c06a18b5e3b1e83d49e" Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.542183 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac423d09dad304a98658b773e736b0d212a9fc7c61136c06a18b5e3b1e83d49e"} err="failed to get container status \"ac423d09dad304a98658b773e736b0d212a9fc7c61136c06a18b5e3b1e83d49e\": rpc error: code = NotFound desc = could not find container \"ac423d09dad304a98658b773e736b0d212a9fc7c61136c06a18b5e3b1e83d49e\": container with ID starting with ac423d09dad304a98658b773e736b0d212a9fc7c61136c06a18b5e3b1e83d49e not found: ID does not exist" Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.542207 4828 scope.go:117] "RemoveContainer" containerID="cb293c341c98f7bb0ea144a5422c436c2724fbbe931cb1686b1ec7e4306513eb" Dec 10 19:56:09 crc kubenswrapper[4828]: E1210 19:56:09.542536 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb293c341c98f7bb0ea144a5422c436c2724fbbe931cb1686b1ec7e4306513eb\": container with ID starting with cb293c341c98f7bb0ea144a5422c436c2724fbbe931cb1686b1ec7e4306513eb not found: ID does not exist" containerID="cb293c341c98f7bb0ea144a5422c436c2724fbbe931cb1686b1ec7e4306513eb" Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.542575 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb293c341c98f7bb0ea144a5422c436c2724fbbe931cb1686b1ec7e4306513eb"} err="failed to get container status \"cb293c341c98f7bb0ea144a5422c436c2724fbbe931cb1686b1ec7e4306513eb\": rpc error: code = NotFound desc = could not find container \"cb293c341c98f7bb0ea144a5422c436c2724fbbe931cb1686b1ec7e4306513eb\": container with ID starting with cb293c341c98f7bb0ea144a5422c436c2724fbbe931cb1686b1ec7e4306513eb not found: ID does not exist" Dec 10 19:56:09 crc kubenswrapper[4828]: I1210 19:56:09.802907 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2814683-d1f3-4307-a95c-94eefe173bf8" path="/var/lib/kubelet/pods/b2814683-d1f3-4307-a95c-94eefe173bf8/volumes" Dec 10 19:56:15 crc kubenswrapper[4828]: I1210 19:56:15.789770 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:56:15 crc kubenswrapper[4828]: E1210 19:56:15.790783 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:56:29 crc kubenswrapper[4828]: I1210 19:56:29.797419 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:56:29 crc kubenswrapper[4828]: E1210 19:56:29.798714 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:56:41 crc kubenswrapper[4828]: I1210 19:56:41.789714 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:56:41 crc kubenswrapper[4828]: E1210 19:56:41.790575 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:56:52 crc kubenswrapper[4828]: I1210 19:56:52.789603 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:56:52 crc kubenswrapper[4828]: E1210 19:56:52.790474 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:57:03 crc kubenswrapper[4828]: I1210 19:57:03.789578 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:57:03 crc kubenswrapper[4828]: E1210 19:57:03.790872 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:57:17 crc kubenswrapper[4828]: E1210 19:57:17.029740 4828 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.214:45812->38.102.83.214:46219: write tcp 38.102.83.214:45812->38.102.83.214:46219: write: connection reset by peer Dec 10 19:57:18 crc kubenswrapper[4828]: I1210 19:57:18.789506 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:57:18 crc kubenswrapper[4828]: E1210 19:57:18.790227 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:57:29 crc kubenswrapper[4828]: I1210 19:57:29.802997 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:57:29 crc kubenswrapper[4828]: E1210 19:57:29.803814 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:57:44 crc kubenswrapper[4828]: I1210 19:57:44.788723 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:57:44 crc kubenswrapper[4828]: E1210 19:57:44.790045 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:57:55 crc kubenswrapper[4828]: I1210 19:57:55.789117 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:57:55 crc kubenswrapper[4828]: E1210 19:57:55.789790 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:58:09 crc kubenswrapper[4828]: I1210 19:58:09.800568 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:58:09 crc kubenswrapper[4828]: E1210 19:58:09.801275 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:58:22 crc kubenswrapper[4828]: I1210 19:58:22.789112 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:58:22 crc kubenswrapper[4828]: E1210 19:58:22.789731 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:58:37 crc kubenswrapper[4828]: I1210 19:58:37.790282 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:58:37 crc kubenswrapper[4828]: E1210 19:58:37.791025 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:58:51 crc kubenswrapper[4828]: I1210 19:58:51.789812 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:58:51 crc kubenswrapper[4828]: E1210 19:58:51.790584 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:59:05 crc kubenswrapper[4828]: I1210 19:59:05.789665 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:59:05 crc kubenswrapper[4828]: E1210 19:59:05.790548 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:59:19 crc kubenswrapper[4828]: I1210 19:59:19.798939 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:59:19 crc kubenswrapper[4828]: E1210 19:59:19.799982 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:59:30 crc kubenswrapper[4828]: I1210 19:59:30.790026 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:59:30 crc kubenswrapper[4828]: E1210 19:59:30.790834 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:59:44 crc kubenswrapper[4828]: I1210 19:59:44.788886 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:59:44 crc kubenswrapper[4828]: E1210 19:59:44.789676 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 19:59:57 crc kubenswrapper[4828]: I1210 19:59:57.789551 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 19:59:57 crc kubenswrapper[4828]: E1210 19:59:57.790235 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.173836 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc"] Dec 10 20:00:00 crc kubenswrapper[4828]: E1210 20:00:00.174905 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2814683-d1f3-4307-a95c-94eefe173bf8" containerName="extract-content" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.174925 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2814683-d1f3-4307-a95c-94eefe173bf8" containerName="extract-content" Dec 10 20:00:00 crc kubenswrapper[4828]: E1210 20:00:00.174969 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ced0629-eb0b-4acd-b2e7-d521f29536f9" containerName="extract-content" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.174977 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ced0629-eb0b-4acd-b2e7-d521f29536f9" containerName="extract-content" Dec 10 20:00:00 crc kubenswrapper[4828]: E1210 20:00:00.174990 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ced0629-eb0b-4acd-b2e7-d521f29536f9" containerName="registry-server" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.175000 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ced0629-eb0b-4acd-b2e7-d521f29536f9" containerName="registry-server" Dec 10 20:00:00 crc kubenswrapper[4828]: E1210 20:00:00.175026 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ced0629-eb0b-4acd-b2e7-d521f29536f9" containerName="extract-utilities" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.175034 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ced0629-eb0b-4acd-b2e7-d521f29536f9" containerName="extract-utilities" Dec 10 20:00:00 crc kubenswrapper[4828]: E1210 20:00:00.175053 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2814683-d1f3-4307-a95c-94eefe173bf8" containerName="registry-server" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.175060 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2814683-d1f3-4307-a95c-94eefe173bf8" containerName="registry-server" Dec 10 20:00:00 crc kubenswrapper[4828]: E1210 20:00:00.175081 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2814683-d1f3-4307-a95c-94eefe173bf8" containerName="extract-utilities" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.175089 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2814683-d1f3-4307-a95c-94eefe173bf8" containerName="extract-utilities" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.175372 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ced0629-eb0b-4acd-b2e7-d521f29536f9" containerName="registry-server" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.175402 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2814683-d1f3-4307-a95c-94eefe173bf8" containerName="registry-server" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.176374 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.178740 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.179039 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.191206 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc"] Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.205026 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9134cad-1cb2-4a16-b5c2-616e11935d72-config-volume\") pod \"collect-profiles-29423280-b4qdc\" (UID: \"f9134cad-1cb2-4a16-b5c2-616e11935d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.205137 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9x5p\" (UniqueName: \"kubernetes.io/projected/f9134cad-1cb2-4a16-b5c2-616e11935d72-kube-api-access-q9x5p\") pod \"collect-profiles-29423280-b4qdc\" (UID: \"f9134cad-1cb2-4a16-b5c2-616e11935d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.205265 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9134cad-1cb2-4a16-b5c2-616e11935d72-secret-volume\") pod \"collect-profiles-29423280-b4qdc\" (UID: \"f9134cad-1cb2-4a16-b5c2-616e11935d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.308029 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9134cad-1cb2-4a16-b5c2-616e11935d72-config-volume\") pod \"collect-profiles-29423280-b4qdc\" (UID: \"f9134cad-1cb2-4a16-b5c2-616e11935d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.308377 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9x5p\" (UniqueName: \"kubernetes.io/projected/f9134cad-1cb2-4a16-b5c2-616e11935d72-kube-api-access-q9x5p\") pod \"collect-profiles-29423280-b4qdc\" (UID: \"f9134cad-1cb2-4a16-b5c2-616e11935d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.308416 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9134cad-1cb2-4a16-b5c2-616e11935d72-secret-volume\") pod \"collect-profiles-29423280-b4qdc\" (UID: \"f9134cad-1cb2-4a16-b5c2-616e11935d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.309121 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9134cad-1cb2-4a16-b5c2-616e11935d72-config-volume\") pod \"collect-profiles-29423280-b4qdc\" (UID: \"f9134cad-1cb2-4a16-b5c2-616e11935d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.314914 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9134cad-1cb2-4a16-b5c2-616e11935d72-secret-volume\") pod \"collect-profiles-29423280-b4qdc\" (UID: \"f9134cad-1cb2-4a16-b5c2-616e11935d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.327418 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9x5p\" (UniqueName: \"kubernetes.io/projected/f9134cad-1cb2-4a16-b5c2-616e11935d72-kube-api-access-q9x5p\") pod \"collect-profiles-29423280-b4qdc\" (UID: \"f9134cad-1cb2-4a16-b5c2-616e11935d72\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc" Dec 10 20:00:00 crc kubenswrapper[4828]: I1210 20:00:00.500115 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc" Dec 10 20:00:01 crc kubenswrapper[4828]: I1210 20:00:01.014294 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc"] Dec 10 20:00:01 crc kubenswrapper[4828]: I1210 20:00:01.963399 4828 generic.go:334] "Generic (PLEG): container finished" podID="f9134cad-1cb2-4a16-b5c2-616e11935d72" containerID="ff4f51293016314e7103c90193feff376ef57d232da1d44a15cc12e24ed8fa54" exitCode=0 Dec 10 20:00:01 crc kubenswrapper[4828]: I1210 20:00:01.963482 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc" event={"ID":"f9134cad-1cb2-4a16-b5c2-616e11935d72","Type":"ContainerDied","Data":"ff4f51293016314e7103c90193feff376ef57d232da1d44a15cc12e24ed8fa54"} Dec 10 20:00:01 crc kubenswrapper[4828]: I1210 20:00:01.963663 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc" event={"ID":"f9134cad-1cb2-4a16-b5c2-616e11935d72","Type":"ContainerStarted","Data":"7acac8dba5e0e50a23db0b31da0321df29b3ce308b080c44c83c464266f2da0d"} Dec 10 20:00:03 crc kubenswrapper[4828]: I1210 20:00:03.528963 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc" Dec 10 20:00:03 crc kubenswrapper[4828]: I1210 20:00:03.587517 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9x5p\" (UniqueName: \"kubernetes.io/projected/f9134cad-1cb2-4a16-b5c2-616e11935d72-kube-api-access-q9x5p\") pod \"f9134cad-1cb2-4a16-b5c2-616e11935d72\" (UID: \"f9134cad-1cb2-4a16-b5c2-616e11935d72\") " Dec 10 20:00:03 crc kubenswrapper[4828]: I1210 20:00:03.588266 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9134cad-1cb2-4a16-b5c2-616e11935d72-secret-volume\") pod \"f9134cad-1cb2-4a16-b5c2-616e11935d72\" (UID: \"f9134cad-1cb2-4a16-b5c2-616e11935d72\") " Dec 10 20:00:03 crc kubenswrapper[4828]: I1210 20:00:03.588336 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9134cad-1cb2-4a16-b5c2-616e11935d72-config-volume\") pod \"f9134cad-1cb2-4a16-b5c2-616e11935d72\" (UID: \"f9134cad-1cb2-4a16-b5c2-616e11935d72\") " Dec 10 20:00:03 crc kubenswrapper[4828]: I1210 20:00:03.589528 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9134cad-1cb2-4a16-b5c2-616e11935d72-config-volume" (OuterVolumeSpecName: "config-volume") pod "f9134cad-1cb2-4a16-b5c2-616e11935d72" (UID: "f9134cad-1cb2-4a16-b5c2-616e11935d72"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 20:00:03 crc kubenswrapper[4828]: I1210 20:00:03.595431 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9134cad-1cb2-4a16-b5c2-616e11935d72-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f9134cad-1cb2-4a16-b5c2-616e11935d72" (UID: "f9134cad-1cb2-4a16-b5c2-616e11935d72"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 20:00:03 crc kubenswrapper[4828]: I1210 20:00:03.595641 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9134cad-1cb2-4a16-b5c2-616e11935d72-kube-api-access-q9x5p" (OuterVolumeSpecName: "kube-api-access-q9x5p") pod "f9134cad-1cb2-4a16-b5c2-616e11935d72" (UID: "f9134cad-1cb2-4a16-b5c2-616e11935d72"). InnerVolumeSpecName "kube-api-access-q9x5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:00:03 crc kubenswrapper[4828]: I1210 20:00:03.691147 4828 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f9134cad-1cb2-4a16-b5c2-616e11935d72-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 20:00:03 crc kubenswrapper[4828]: I1210 20:00:03.691226 4828 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f9134cad-1cb2-4a16-b5c2-616e11935d72-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 20:00:03 crc kubenswrapper[4828]: I1210 20:00:03.691242 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9x5p\" (UniqueName: \"kubernetes.io/projected/f9134cad-1cb2-4a16-b5c2-616e11935d72-kube-api-access-q9x5p\") on node \"crc\" DevicePath \"\"" Dec 10 20:00:04 crc kubenswrapper[4828]: I1210 20:00:04.000776 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc" event={"ID":"f9134cad-1cb2-4a16-b5c2-616e11935d72","Type":"ContainerDied","Data":"7acac8dba5e0e50a23db0b31da0321df29b3ce308b080c44c83c464266f2da0d"} Dec 10 20:00:04 crc kubenswrapper[4828]: I1210 20:00:04.000873 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7acac8dba5e0e50a23db0b31da0321df29b3ce308b080c44c83c464266f2da0d" Dec 10 20:00:04 crc kubenswrapper[4828]: I1210 20:00:04.000940 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423280-b4qdc" Dec 10 20:00:04 crc kubenswrapper[4828]: I1210 20:00:04.612019 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k"] Dec 10 20:00:04 crc kubenswrapper[4828]: I1210 20:00:04.624401 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423235-pg58k"] Dec 10 20:00:05 crc kubenswrapper[4828]: I1210 20:00:05.805765 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13848851-6a8f-48ce-a64e-870a6d47993d" path="/var/lib/kubelet/pods/13848851-6a8f-48ce-a64e-870a6d47993d/volumes" Dec 10 20:00:10 crc kubenswrapper[4828]: I1210 20:00:10.788960 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 20:00:10 crc kubenswrapper[4828]: E1210 20:00:10.789982 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:00:22 crc kubenswrapper[4828]: I1210 20:00:22.789311 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 20:00:23 crc kubenswrapper[4828]: I1210 20:00:23.198430 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"db5d83b0d9b828f200e2cc4f0121342bbc11c64408cff1c60141175d4820ef05"} Dec 10 20:00:42 crc kubenswrapper[4828]: I1210 20:00:42.650708 4828 scope.go:117] "RemoveContainer" containerID="f227a7ac24dbd3b95a332f31c411345447e280b0d5ac94889b590869ae2e8c0b" Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.149672 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29423281-rkwnp"] Dec 10 20:01:00 crc kubenswrapper[4828]: E1210 20:01:00.151661 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9134cad-1cb2-4a16-b5c2-616e11935d72" containerName="collect-profiles" Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.151747 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9134cad-1cb2-4a16-b5c2-616e11935d72" containerName="collect-profiles" Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.152086 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9134cad-1cb2-4a16-b5c2-616e11935d72" containerName="collect-profiles" Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.153741 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29423281-rkwnp" Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.167859 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29423281-rkwnp"] Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.268312 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-config-data\") pod \"keystone-cron-29423281-rkwnp\" (UID: \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\") " pod="openstack/keystone-cron-29423281-rkwnp" Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.268364 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-987g4\" (UniqueName: \"kubernetes.io/projected/40346cd5-252e-4737-a7ff-8dacb87c3a7f-kube-api-access-987g4\") pod \"keystone-cron-29423281-rkwnp\" (UID: \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\") " pod="openstack/keystone-cron-29423281-rkwnp" Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.268663 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-combined-ca-bundle\") pod \"keystone-cron-29423281-rkwnp\" (UID: \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\") " pod="openstack/keystone-cron-29423281-rkwnp" Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.268973 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-fernet-keys\") pod \"keystone-cron-29423281-rkwnp\" (UID: \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\") " pod="openstack/keystone-cron-29423281-rkwnp" Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.371325 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-combined-ca-bundle\") pod \"keystone-cron-29423281-rkwnp\" (UID: \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\") " pod="openstack/keystone-cron-29423281-rkwnp" Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.371433 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-fernet-keys\") pod \"keystone-cron-29423281-rkwnp\" (UID: \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\") " pod="openstack/keystone-cron-29423281-rkwnp" Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.371655 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-config-data\") pod \"keystone-cron-29423281-rkwnp\" (UID: \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\") " pod="openstack/keystone-cron-29423281-rkwnp" Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.371696 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-987g4\" (UniqueName: \"kubernetes.io/projected/40346cd5-252e-4737-a7ff-8dacb87c3a7f-kube-api-access-987g4\") pod \"keystone-cron-29423281-rkwnp\" (UID: \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\") " pod="openstack/keystone-cron-29423281-rkwnp" Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.379022 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-fernet-keys\") pod \"keystone-cron-29423281-rkwnp\" (UID: \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\") " pod="openstack/keystone-cron-29423281-rkwnp" Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.379351 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-config-data\") pod \"keystone-cron-29423281-rkwnp\" (UID: \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\") " pod="openstack/keystone-cron-29423281-rkwnp" Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.379578 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-combined-ca-bundle\") pod \"keystone-cron-29423281-rkwnp\" (UID: \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\") " pod="openstack/keystone-cron-29423281-rkwnp" Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.394859 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-987g4\" (UniqueName: \"kubernetes.io/projected/40346cd5-252e-4737-a7ff-8dacb87c3a7f-kube-api-access-987g4\") pod \"keystone-cron-29423281-rkwnp\" (UID: \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\") " pod="openstack/keystone-cron-29423281-rkwnp" Dec 10 20:01:00 crc kubenswrapper[4828]: I1210 20:01:00.476711 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29423281-rkwnp" Dec 10 20:01:01 crc kubenswrapper[4828]: I1210 20:01:01.003976 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29423281-rkwnp"] Dec 10 20:01:01 crc kubenswrapper[4828]: I1210 20:01:01.586426 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29423281-rkwnp" event={"ID":"40346cd5-252e-4737-a7ff-8dacb87c3a7f","Type":"ContainerStarted","Data":"be28f97e40c69023a1e021aa4c8655cb4a93009bd5bf147ed685c8950c1306dc"} Dec 10 20:01:01 crc kubenswrapper[4828]: I1210 20:01:01.586666 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29423281-rkwnp" event={"ID":"40346cd5-252e-4737-a7ff-8dacb87c3a7f","Type":"ContainerStarted","Data":"632f7f3def1286da17f85e7b07a310c1ad451743dc2e9280434388b9f1d658b2"} Dec 10 20:01:01 crc kubenswrapper[4828]: I1210 20:01:01.615500 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29423281-rkwnp" podStartSLOduration=1.6154736 podStartE2EDuration="1.6154736s" podCreationTimestamp="2025-12-10 20:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 20:01:01.609195003 +0000 UTC m=+3942.119806028" watchObservedRunningTime="2025-12-10 20:01:01.6154736 +0000 UTC m=+3942.126084605" Dec 10 20:01:04 crc kubenswrapper[4828]: I1210 20:01:04.622850 4828 generic.go:334] "Generic (PLEG): container finished" podID="40346cd5-252e-4737-a7ff-8dacb87c3a7f" containerID="be28f97e40c69023a1e021aa4c8655cb4a93009bd5bf147ed685c8950c1306dc" exitCode=0 Dec 10 20:01:04 crc kubenswrapper[4828]: I1210 20:01:04.622946 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29423281-rkwnp" event={"ID":"40346cd5-252e-4737-a7ff-8dacb87c3a7f","Type":"ContainerDied","Data":"be28f97e40c69023a1e021aa4c8655cb4a93009bd5bf147ed685c8950c1306dc"} Dec 10 20:01:06 crc kubenswrapper[4828]: I1210 20:01:06.112361 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29423281-rkwnp" Dec 10 20:01:06 crc kubenswrapper[4828]: I1210 20:01:06.211231 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-fernet-keys\") pod \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\" (UID: \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\") " Dec 10 20:01:06 crc kubenswrapper[4828]: I1210 20:01:06.211287 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-config-data\") pod \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\" (UID: \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\") " Dec 10 20:01:06 crc kubenswrapper[4828]: I1210 20:01:06.211533 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-987g4\" (UniqueName: \"kubernetes.io/projected/40346cd5-252e-4737-a7ff-8dacb87c3a7f-kube-api-access-987g4\") pod \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\" (UID: \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\") " Dec 10 20:01:06 crc kubenswrapper[4828]: I1210 20:01:06.211610 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-combined-ca-bundle\") pod \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\" (UID: \"40346cd5-252e-4737-a7ff-8dacb87c3a7f\") " Dec 10 20:01:06 crc kubenswrapper[4828]: I1210 20:01:06.216624 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "40346cd5-252e-4737-a7ff-8dacb87c3a7f" (UID: "40346cd5-252e-4737-a7ff-8dacb87c3a7f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 20:01:06 crc kubenswrapper[4828]: I1210 20:01:06.221432 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40346cd5-252e-4737-a7ff-8dacb87c3a7f-kube-api-access-987g4" (OuterVolumeSpecName: "kube-api-access-987g4") pod "40346cd5-252e-4737-a7ff-8dacb87c3a7f" (UID: "40346cd5-252e-4737-a7ff-8dacb87c3a7f"). InnerVolumeSpecName "kube-api-access-987g4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:01:06 crc kubenswrapper[4828]: I1210 20:01:06.255375 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40346cd5-252e-4737-a7ff-8dacb87c3a7f" (UID: "40346cd5-252e-4737-a7ff-8dacb87c3a7f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 20:01:06 crc kubenswrapper[4828]: I1210 20:01:06.291050 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-config-data" (OuterVolumeSpecName: "config-data") pod "40346cd5-252e-4737-a7ff-8dacb87c3a7f" (UID: "40346cd5-252e-4737-a7ff-8dacb87c3a7f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 20:01:06 crc kubenswrapper[4828]: I1210 20:01:06.314590 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-987g4\" (UniqueName: \"kubernetes.io/projected/40346cd5-252e-4737-a7ff-8dacb87c3a7f-kube-api-access-987g4\") on node \"crc\" DevicePath \"\"" Dec 10 20:01:06 crc kubenswrapper[4828]: I1210 20:01:06.314641 4828 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 10 20:01:06 crc kubenswrapper[4828]: I1210 20:01:06.314657 4828 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 10 20:01:06 crc kubenswrapper[4828]: I1210 20:01:06.314671 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40346cd5-252e-4737-a7ff-8dacb87c3a7f-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 20:01:06 crc kubenswrapper[4828]: I1210 20:01:06.645323 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29423281-rkwnp" Dec 10 20:01:06 crc kubenswrapper[4828]: I1210 20:01:06.645321 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29423281-rkwnp" event={"ID":"40346cd5-252e-4737-a7ff-8dacb87c3a7f","Type":"ContainerDied","Data":"632f7f3def1286da17f85e7b07a310c1ad451743dc2e9280434388b9f1d658b2"} Dec 10 20:01:06 crc kubenswrapper[4828]: I1210 20:01:06.645436 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="632f7f3def1286da17f85e7b07a310c1ad451743dc2e9280434388b9f1d658b2" Dec 10 20:02:51 crc kubenswrapper[4828]: I1210 20:02:51.230284 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:02:51 crc kubenswrapper[4828]: I1210 20:02:51.231037 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:02:52 crc kubenswrapper[4828]: E1210 20:02:52.513083 4828 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.214:47792->38.102.83.214:46219: read tcp 38.102.83.214:47792->38.102.83.214:46219: read: connection reset by peer Dec 10 20:03:21 crc kubenswrapper[4828]: I1210 20:03:21.230506 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:03:21 crc kubenswrapper[4828]: I1210 20:03:21.231158 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:03:34 crc kubenswrapper[4828]: I1210 20:03:34.070309 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f2p98"] Dec 10 20:03:34 crc kubenswrapper[4828]: E1210 20:03:34.073391 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40346cd5-252e-4737-a7ff-8dacb87c3a7f" containerName="keystone-cron" Dec 10 20:03:34 crc kubenswrapper[4828]: I1210 20:03:34.074365 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="40346cd5-252e-4737-a7ff-8dacb87c3a7f" containerName="keystone-cron" Dec 10 20:03:34 crc kubenswrapper[4828]: I1210 20:03:34.075038 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="40346cd5-252e-4737-a7ff-8dacb87c3a7f" containerName="keystone-cron" Dec 10 20:03:34 crc kubenswrapper[4828]: I1210 20:03:34.078269 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f2p98" Dec 10 20:03:34 crc kubenswrapper[4828]: I1210 20:03:34.081968 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f2p98"] Dec 10 20:03:34 crc kubenswrapper[4828]: I1210 20:03:34.132566 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dcf252d-ed7e-4dff-ad86-18de83547fbf-utilities\") pod \"redhat-operators-f2p98\" (UID: \"0dcf252d-ed7e-4dff-ad86-18de83547fbf\") " pod="openshift-marketplace/redhat-operators-f2p98" Dec 10 20:03:34 crc kubenswrapper[4828]: I1210 20:03:34.132653 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dbhb\" (UniqueName: \"kubernetes.io/projected/0dcf252d-ed7e-4dff-ad86-18de83547fbf-kube-api-access-2dbhb\") pod \"redhat-operators-f2p98\" (UID: \"0dcf252d-ed7e-4dff-ad86-18de83547fbf\") " pod="openshift-marketplace/redhat-operators-f2p98" Dec 10 20:03:34 crc kubenswrapper[4828]: I1210 20:03:34.132950 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dcf252d-ed7e-4dff-ad86-18de83547fbf-catalog-content\") pod \"redhat-operators-f2p98\" (UID: \"0dcf252d-ed7e-4dff-ad86-18de83547fbf\") " pod="openshift-marketplace/redhat-operators-f2p98" Dec 10 20:03:34 crc kubenswrapper[4828]: I1210 20:03:34.239409 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dcf252d-ed7e-4dff-ad86-18de83547fbf-utilities\") pod \"redhat-operators-f2p98\" (UID: \"0dcf252d-ed7e-4dff-ad86-18de83547fbf\") " pod="openshift-marketplace/redhat-operators-f2p98" Dec 10 20:03:34 crc kubenswrapper[4828]: I1210 20:03:34.239498 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dbhb\" (UniqueName: \"kubernetes.io/projected/0dcf252d-ed7e-4dff-ad86-18de83547fbf-kube-api-access-2dbhb\") pod \"redhat-operators-f2p98\" (UID: \"0dcf252d-ed7e-4dff-ad86-18de83547fbf\") " pod="openshift-marketplace/redhat-operators-f2p98" Dec 10 20:03:34 crc kubenswrapper[4828]: I1210 20:03:34.239553 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dcf252d-ed7e-4dff-ad86-18de83547fbf-catalog-content\") pod \"redhat-operators-f2p98\" (UID: \"0dcf252d-ed7e-4dff-ad86-18de83547fbf\") " pod="openshift-marketplace/redhat-operators-f2p98" Dec 10 20:03:34 crc kubenswrapper[4828]: I1210 20:03:34.240190 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dcf252d-ed7e-4dff-ad86-18de83547fbf-catalog-content\") pod \"redhat-operators-f2p98\" (UID: \"0dcf252d-ed7e-4dff-ad86-18de83547fbf\") " pod="openshift-marketplace/redhat-operators-f2p98" Dec 10 20:03:34 crc kubenswrapper[4828]: I1210 20:03:34.240377 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dcf252d-ed7e-4dff-ad86-18de83547fbf-utilities\") pod \"redhat-operators-f2p98\" (UID: \"0dcf252d-ed7e-4dff-ad86-18de83547fbf\") " pod="openshift-marketplace/redhat-operators-f2p98" Dec 10 20:03:34 crc kubenswrapper[4828]: I1210 20:03:34.264978 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dbhb\" (UniqueName: \"kubernetes.io/projected/0dcf252d-ed7e-4dff-ad86-18de83547fbf-kube-api-access-2dbhb\") pod \"redhat-operators-f2p98\" (UID: \"0dcf252d-ed7e-4dff-ad86-18de83547fbf\") " pod="openshift-marketplace/redhat-operators-f2p98" Dec 10 20:03:34 crc kubenswrapper[4828]: I1210 20:03:34.406120 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f2p98" Dec 10 20:03:35 crc kubenswrapper[4828]: I1210 20:03:35.479677 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f2p98"] Dec 10 20:03:36 crc kubenswrapper[4828]: I1210 20:03:36.280612 4828 generic.go:334] "Generic (PLEG): container finished" podID="0dcf252d-ed7e-4dff-ad86-18de83547fbf" containerID="42f171f30531b60b7ba9951b67f76e7756704b291dd9e11933736f5966c45eaf" exitCode=0 Dec 10 20:03:36 crc kubenswrapper[4828]: I1210 20:03:36.280667 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f2p98" event={"ID":"0dcf252d-ed7e-4dff-ad86-18de83547fbf","Type":"ContainerDied","Data":"42f171f30531b60b7ba9951b67f76e7756704b291dd9e11933736f5966c45eaf"} Dec 10 20:03:36 crc kubenswrapper[4828]: I1210 20:03:36.280909 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f2p98" event={"ID":"0dcf252d-ed7e-4dff-ad86-18de83547fbf","Type":"ContainerStarted","Data":"aa96dde05b002e5e5ab318c370917ec9248f6fa3086acc8bdaef35b6a16ec540"} Dec 10 20:03:36 crc kubenswrapper[4828]: I1210 20:03:36.284369 4828 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 20:03:38 crc kubenswrapper[4828]: I1210 20:03:38.305699 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f2p98" event={"ID":"0dcf252d-ed7e-4dff-ad86-18de83547fbf","Type":"ContainerStarted","Data":"46184b2a26c3ef5fc2619f5906c01b330d3235bb8ca56ddb92d7bac0a8790e76"} Dec 10 20:03:41 crc kubenswrapper[4828]: I1210 20:03:41.358249 4828 generic.go:334] "Generic (PLEG): container finished" podID="0dcf252d-ed7e-4dff-ad86-18de83547fbf" containerID="46184b2a26c3ef5fc2619f5906c01b330d3235bb8ca56ddb92d7bac0a8790e76" exitCode=0 Dec 10 20:03:41 crc kubenswrapper[4828]: I1210 20:03:41.358339 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f2p98" event={"ID":"0dcf252d-ed7e-4dff-ad86-18de83547fbf","Type":"ContainerDied","Data":"46184b2a26c3ef5fc2619f5906c01b330d3235bb8ca56ddb92d7bac0a8790e76"} Dec 10 20:03:43 crc kubenswrapper[4828]: I1210 20:03:43.380460 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f2p98" event={"ID":"0dcf252d-ed7e-4dff-ad86-18de83547fbf","Type":"ContainerStarted","Data":"e4e39000929330927554a4e8771fce03d4a2104ce408daae52d2452396a9d39b"} Dec 10 20:03:43 crc kubenswrapper[4828]: I1210 20:03:43.416380 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f2p98" podStartSLOduration=3.275518439 podStartE2EDuration="9.4163636s" podCreationTimestamp="2025-12-10 20:03:34 +0000 UTC" firstStartedPulling="2025-12-10 20:03:36.28413666 +0000 UTC m=+4096.794747665" lastFinishedPulling="2025-12-10 20:03:42.424981821 +0000 UTC m=+4102.935592826" observedRunningTime="2025-12-10 20:03:43.412355357 +0000 UTC m=+4103.922966382" watchObservedRunningTime="2025-12-10 20:03:43.4163636 +0000 UTC m=+4103.926974605" Dec 10 20:03:44 crc kubenswrapper[4828]: I1210 20:03:44.407628 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f2p98" Dec 10 20:03:44 crc kubenswrapper[4828]: I1210 20:03:44.408032 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f2p98" Dec 10 20:03:45 crc kubenswrapper[4828]: I1210 20:03:45.459879 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-f2p98" podUID="0dcf252d-ed7e-4dff-ad86-18de83547fbf" containerName="registry-server" probeResult="failure" output=< Dec 10 20:03:45 crc kubenswrapper[4828]: timeout: failed to connect service ":50051" within 1s Dec 10 20:03:45 crc kubenswrapper[4828]: > Dec 10 20:03:51 crc kubenswrapper[4828]: I1210 20:03:51.230603 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:03:51 crc kubenswrapper[4828]: I1210 20:03:51.231210 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:03:51 crc kubenswrapper[4828]: I1210 20:03:51.231518 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 20:03:51 crc kubenswrapper[4828]: I1210 20:03:51.232312 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"db5d83b0d9b828f200e2cc4f0121342bbc11c64408cff1c60141175d4820ef05"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 20:03:51 crc kubenswrapper[4828]: I1210 20:03:51.232384 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://db5d83b0d9b828f200e2cc4f0121342bbc11c64408cff1c60141175d4820ef05" gracePeriod=600 Dec 10 20:03:51 crc kubenswrapper[4828]: I1210 20:03:51.483171 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="db5d83b0d9b828f200e2cc4f0121342bbc11c64408cff1c60141175d4820ef05" exitCode=0 Dec 10 20:03:51 crc kubenswrapper[4828]: I1210 20:03:51.483326 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"db5d83b0d9b828f200e2cc4f0121342bbc11c64408cff1c60141175d4820ef05"} Dec 10 20:03:51 crc kubenswrapper[4828]: I1210 20:03:51.483525 4828 scope.go:117] "RemoveContainer" containerID="28410d82f6dc32de6b24682bd316d1b51a998071b293cb8c403f5eacb89783dd" Dec 10 20:03:52 crc kubenswrapper[4828]: I1210 20:03:52.499111 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782"} Dec 10 20:03:54 crc kubenswrapper[4828]: I1210 20:03:54.463718 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f2p98" Dec 10 20:03:54 crc kubenswrapper[4828]: I1210 20:03:54.525263 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f2p98" Dec 10 20:03:55 crc kubenswrapper[4828]: I1210 20:03:55.571403 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f2p98"] Dec 10 20:03:55 crc kubenswrapper[4828]: I1210 20:03:55.572415 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f2p98" podUID="0dcf252d-ed7e-4dff-ad86-18de83547fbf" containerName="registry-server" containerID="cri-o://e4e39000929330927554a4e8771fce03d4a2104ce408daae52d2452396a9d39b" gracePeriod=2 Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.139677 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f2p98" Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.243275 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dcf252d-ed7e-4dff-ad86-18de83547fbf-utilities\") pod \"0dcf252d-ed7e-4dff-ad86-18de83547fbf\" (UID: \"0dcf252d-ed7e-4dff-ad86-18de83547fbf\") " Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.243412 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dbhb\" (UniqueName: \"kubernetes.io/projected/0dcf252d-ed7e-4dff-ad86-18de83547fbf-kube-api-access-2dbhb\") pod \"0dcf252d-ed7e-4dff-ad86-18de83547fbf\" (UID: \"0dcf252d-ed7e-4dff-ad86-18de83547fbf\") " Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.243504 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dcf252d-ed7e-4dff-ad86-18de83547fbf-catalog-content\") pod \"0dcf252d-ed7e-4dff-ad86-18de83547fbf\" (UID: \"0dcf252d-ed7e-4dff-ad86-18de83547fbf\") " Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.244564 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dcf252d-ed7e-4dff-ad86-18de83547fbf-utilities" (OuterVolumeSpecName: "utilities") pod "0dcf252d-ed7e-4dff-ad86-18de83547fbf" (UID: "0dcf252d-ed7e-4dff-ad86-18de83547fbf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.277029 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dcf252d-ed7e-4dff-ad86-18de83547fbf-kube-api-access-2dbhb" (OuterVolumeSpecName: "kube-api-access-2dbhb") pod "0dcf252d-ed7e-4dff-ad86-18de83547fbf" (UID: "0dcf252d-ed7e-4dff-ad86-18de83547fbf"). InnerVolumeSpecName "kube-api-access-2dbhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.346758 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0dcf252d-ed7e-4dff-ad86-18de83547fbf-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.346821 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dbhb\" (UniqueName: \"kubernetes.io/projected/0dcf252d-ed7e-4dff-ad86-18de83547fbf-kube-api-access-2dbhb\") on node \"crc\" DevicePath \"\"" Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.377486 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dcf252d-ed7e-4dff-ad86-18de83547fbf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0dcf252d-ed7e-4dff-ad86-18de83547fbf" (UID: "0dcf252d-ed7e-4dff-ad86-18de83547fbf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.449437 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0dcf252d-ed7e-4dff-ad86-18de83547fbf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.548257 4828 generic.go:334] "Generic (PLEG): container finished" podID="0dcf252d-ed7e-4dff-ad86-18de83547fbf" containerID="e4e39000929330927554a4e8771fce03d4a2104ce408daae52d2452396a9d39b" exitCode=0 Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.548316 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f2p98" event={"ID":"0dcf252d-ed7e-4dff-ad86-18de83547fbf","Type":"ContainerDied","Data":"e4e39000929330927554a4e8771fce03d4a2104ce408daae52d2452396a9d39b"} Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.548343 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f2p98" event={"ID":"0dcf252d-ed7e-4dff-ad86-18de83547fbf","Type":"ContainerDied","Data":"aa96dde05b002e5e5ab318c370917ec9248f6fa3086acc8bdaef35b6a16ec540"} Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.548357 4828 scope.go:117] "RemoveContainer" containerID="e4e39000929330927554a4e8771fce03d4a2104ce408daae52d2452396a9d39b" Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.548511 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f2p98" Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.575220 4828 scope.go:117] "RemoveContainer" containerID="46184b2a26c3ef5fc2619f5906c01b330d3235bb8ca56ddb92d7bac0a8790e76" Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.589746 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f2p98"] Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.600691 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f2p98"] Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.607240 4828 scope.go:117] "RemoveContainer" containerID="42f171f30531b60b7ba9951b67f76e7756704b291dd9e11933736f5966c45eaf" Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.657434 4828 scope.go:117] "RemoveContainer" containerID="e4e39000929330927554a4e8771fce03d4a2104ce408daae52d2452396a9d39b" Dec 10 20:03:56 crc kubenswrapper[4828]: E1210 20:03:56.657870 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4e39000929330927554a4e8771fce03d4a2104ce408daae52d2452396a9d39b\": container with ID starting with e4e39000929330927554a4e8771fce03d4a2104ce408daae52d2452396a9d39b not found: ID does not exist" containerID="e4e39000929330927554a4e8771fce03d4a2104ce408daae52d2452396a9d39b" Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.657903 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4e39000929330927554a4e8771fce03d4a2104ce408daae52d2452396a9d39b"} err="failed to get container status \"e4e39000929330927554a4e8771fce03d4a2104ce408daae52d2452396a9d39b\": rpc error: code = NotFound desc = could not find container \"e4e39000929330927554a4e8771fce03d4a2104ce408daae52d2452396a9d39b\": container with ID starting with e4e39000929330927554a4e8771fce03d4a2104ce408daae52d2452396a9d39b not found: ID does not exist" Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.657922 4828 scope.go:117] "RemoveContainer" containerID="46184b2a26c3ef5fc2619f5906c01b330d3235bb8ca56ddb92d7bac0a8790e76" Dec 10 20:03:56 crc kubenswrapper[4828]: E1210 20:03:56.659723 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46184b2a26c3ef5fc2619f5906c01b330d3235bb8ca56ddb92d7bac0a8790e76\": container with ID starting with 46184b2a26c3ef5fc2619f5906c01b330d3235bb8ca56ddb92d7bac0a8790e76 not found: ID does not exist" containerID="46184b2a26c3ef5fc2619f5906c01b330d3235bb8ca56ddb92d7bac0a8790e76" Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.659767 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46184b2a26c3ef5fc2619f5906c01b330d3235bb8ca56ddb92d7bac0a8790e76"} err="failed to get container status \"46184b2a26c3ef5fc2619f5906c01b330d3235bb8ca56ddb92d7bac0a8790e76\": rpc error: code = NotFound desc = could not find container \"46184b2a26c3ef5fc2619f5906c01b330d3235bb8ca56ddb92d7bac0a8790e76\": container with ID starting with 46184b2a26c3ef5fc2619f5906c01b330d3235bb8ca56ddb92d7bac0a8790e76 not found: ID does not exist" Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.659807 4828 scope.go:117] "RemoveContainer" containerID="42f171f30531b60b7ba9951b67f76e7756704b291dd9e11933736f5966c45eaf" Dec 10 20:03:56 crc kubenswrapper[4828]: E1210 20:03:56.660092 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42f171f30531b60b7ba9951b67f76e7756704b291dd9e11933736f5966c45eaf\": container with ID starting with 42f171f30531b60b7ba9951b67f76e7756704b291dd9e11933736f5966c45eaf not found: ID does not exist" containerID="42f171f30531b60b7ba9951b67f76e7756704b291dd9e11933736f5966c45eaf" Dec 10 20:03:56 crc kubenswrapper[4828]: I1210 20:03:56.660125 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42f171f30531b60b7ba9951b67f76e7756704b291dd9e11933736f5966c45eaf"} err="failed to get container status \"42f171f30531b60b7ba9951b67f76e7756704b291dd9e11933736f5966c45eaf\": rpc error: code = NotFound desc = could not find container \"42f171f30531b60b7ba9951b67f76e7756704b291dd9e11933736f5966c45eaf\": container with ID starting with 42f171f30531b60b7ba9951b67f76e7756704b291dd9e11933736f5966c45eaf not found: ID does not exist" Dec 10 20:03:57 crc kubenswrapper[4828]: I1210 20:03:57.801657 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dcf252d-ed7e-4dff-ad86-18de83547fbf" path="/var/lib/kubelet/pods/0dcf252d-ed7e-4dff-ad86-18de83547fbf/volumes" Dec 10 20:05:51 crc kubenswrapper[4828]: I1210 20:05:51.230456 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:05:51 crc kubenswrapper[4828]: I1210 20:05:51.231481 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:06:01 crc kubenswrapper[4828]: I1210 20:06:01.900688 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kzvrf"] Dec 10 20:06:01 crc kubenswrapper[4828]: E1210 20:06:01.901792 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dcf252d-ed7e-4dff-ad86-18de83547fbf" containerName="registry-server" Dec 10 20:06:01 crc kubenswrapper[4828]: I1210 20:06:01.901828 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dcf252d-ed7e-4dff-ad86-18de83547fbf" containerName="registry-server" Dec 10 20:06:01 crc kubenswrapper[4828]: E1210 20:06:01.901856 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dcf252d-ed7e-4dff-ad86-18de83547fbf" containerName="extract-utilities" Dec 10 20:06:01 crc kubenswrapper[4828]: I1210 20:06:01.901864 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dcf252d-ed7e-4dff-ad86-18de83547fbf" containerName="extract-utilities" Dec 10 20:06:01 crc kubenswrapper[4828]: E1210 20:06:01.901877 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dcf252d-ed7e-4dff-ad86-18de83547fbf" containerName="extract-content" Dec 10 20:06:01 crc kubenswrapper[4828]: I1210 20:06:01.901885 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dcf252d-ed7e-4dff-ad86-18de83547fbf" containerName="extract-content" Dec 10 20:06:01 crc kubenswrapper[4828]: I1210 20:06:01.902165 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dcf252d-ed7e-4dff-ad86-18de83547fbf" containerName="registry-server" Dec 10 20:06:01 crc kubenswrapper[4828]: I1210 20:06:01.904969 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kzvrf" Dec 10 20:06:01 crc kubenswrapper[4828]: I1210 20:06:01.957624 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kzvrf"] Dec 10 20:06:02 crc kubenswrapper[4828]: I1210 20:06:02.015044 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-utilities\") pod \"redhat-marketplace-kzvrf\" (UID: \"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa\") " pod="openshift-marketplace/redhat-marketplace-kzvrf" Dec 10 20:06:02 crc kubenswrapper[4828]: I1210 20:06:02.015151 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-catalog-content\") pod \"redhat-marketplace-kzvrf\" (UID: \"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa\") " pod="openshift-marketplace/redhat-marketplace-kzvrf" Dec 10 20:06:02 crc kubenswrapper[4828]: I1210 20:06:02.015391 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ghvc\" (UniqueName: \"kubernetes.io/projected/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-kube-api-access-5ghvc\") pod \"redhat-marketplace-kzvrf\" (UID: \"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa\") " pod="openshift-marketplace/redhat-marketplace-kzvrf" Dec 10 20:06:02 crc kubenswrapper[4828]: I1210 20:06:02.117010 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-utilities\") pod \"redhat-marketplace-kzvrf\" (UID: \"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa\") " pod="openshift-marketplace/redhat-marketplace-kzvrf" Dec 10 20:06:02 crc kubenswrapper[4828]: I1210 20:06:02.117054 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-catalog-content\") pod \"redhat-marketplace-kzvrf\" (UID: \"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa\") " pod="openshift-marketplace/redhat-marketplace-kzvrf" Dec 10 20:06:02 crc kubenswrapper[4828]: I1210 20:06:02.117263 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ghvc\" (UniqueName: \"kubernetes.io/projected/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-kube-api-access-5ghvc\") pod \"redhat-marketplace-kzvrf\" (UID: \"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa\") " pod="openshift-marketplace/redhat-marketplace-kzvrf" Dec 10 20:06:02 crc kubenswrapper[4828]: I1210 20:06:02.117492 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-utilities\") pod \"redhat-marketplace-kzvrf\" (UID: \"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa\") " pod="openshift-marketplace/redhat-marketplace-kzvrf" Dec 10 20:06:02 crc kubenswrapper[4828]: I1210 20:06:02.117763 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-catalog-content\") pod \"redhat-marketplace-kzvrf\" (UID: \"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa\") " pod="openshift-marketplace/redhat-marketplace-kzvrf" Dec 10 20:06:02 crc kubenswrapper[4828]: I1210 20:06:02.154891 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ghvc\" (UniqueName: \"kubernetes.io/projected/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-kube-api-access-5ghvc\") pod \"redhat-marketplace-kzvrf\" (UID: \"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa\") " pod="openshift-marketplace/redhat-marketplace-kzvrf" Dec 10 20:06:02 crc kubenswrapper[4828]: I1210 20:06:02.231854 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kzvrf" Dec 10 20:06:02 crc kubenswrapper[4828]: I1210 20:06:02.786501 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kzvrf"] Dec 10 20:06:04 crc kubenswrapper[4828]: I1210 20:06:04.052076 4828 generic.go:334] "Generic (PLEG): container finished" podID="8fe548f7-e9f5-4bc1-8af3-530d7c5819fa" containerID="10c0cd6d8a3a8db050aa2d7fb26e982780dece019dee3c15551a9822832266f4" exitCode=0 Dec 10 20:06:04 crc kubenswrapper[4828]: I1210 20:06:04.052161 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzvrf" event={"ID":"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa","Type":"ContainerDied","Data":"10c0cd6d8a3a8db050aa2d7fb26e982780dece019dee3c15551a9822832266f4"} Dec 10 20:06:04 crc kubenswrapper[4828]: I1210 20:06:04.054102 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzvrf" event={"ID":"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa","Type":"ContainerStarted","Data":"38dee5e51b136f452218a107783340667824018010f45de93f861dad3c2b37a0"} Dec 10 20:06:08 crc kubenswrapper[4828]: I1210 20:06:08.092460 4828 generic.go:334] "Generic (PLEG): container finished" podID="8fe548f7-e9f5-4bc1-8af3-530d7c5819fa" containerID="40c834d7c379fd9f8bdf5ea05e1e19e7e4bcd9c2972921a2b27342e1e5f31095" exitCode=0 Dec 10 20:06:08 crc kubenswrapper[4828]: I1210 20:06:08.092542 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzvrf" event={"ID":"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa","Type":"ContainerDied","Data":"40c834d7c379fd9f8bdf5ea05e1e19e7e4bcd9c2972921a2b27342e1e5f31095"} Dec 10 20:06:09 crc kubenswrapper[4828]: I1210 20:06:09.105049 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzvrf" event={"ID":"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa","Type":"ContainerStarted","Data":"dc8cd3eda11c220972236f9c1dd8e00f8e558f49b3468168d0f6b8c955a9f81e"} Dec 10 20:06:12 crc kubenswrapper[4828]: I1210 20:06:12.232541 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kzvrf" Dec 10 20:06:12 crc kubenswrapper[4828]: I1210 20:06:12.233440 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kzvrf" Dec 10 20:06:12 crc kubenswrapper[4828]: I1210 20:06:12.293116 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kzvrf" Dec 10 20:06:12 crc kubenswrapper[4828]: I1210 20:06:12.322515 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kzvrf" podStartSLOduration=6.850895734 podStartE2EDuration="11.322496342s" podCreationTimestamp="2025-12-10 20:06:01 +0000 UTC" firstStartedPulling="2025-12-10 20:06:04.055398931 +0000 UTC m=+4244.566009936" lastFinishedPulling="2025-12-10 20:06:08.526999539 +0000 UTC m=+4249.037610544" observedRunningTime="2025-12-10 20:06:09.123466546 +0000 UTC m=+4249.634077551" watchObservedRunningTime="2025-12-10 20:06:12.322496342 +0000 UTC m=+4252.833107347" Dec 10 20:06:13 crc kubenswrapper[4828]: I1210 20:06:13.209829 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kzvrf" Dec 10 20:06:13 crc kubenswrapper[4828]: I1210 20:06:13.275873 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kzvrf"] Dec 10 20:06:15 crc kubenswrapper[4828]: I1210 20:06:15.168529 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kzvrf" podUID="8fe548f7-e9f5-4bc1-8af3-530d7c5819fa" containerName="registry-server" containerID="cri-o://dc8cd3eda11c220972236f9c1dd8e00f8e558f49b3468168d0f6b8c955a9f81e" gracePeriod=2 Dec 10 20:06:15 crc kubenswrapper[4828]: I1210 20:06:15.700746 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kzvrf" Dec 10 20:06:15 crc kubenswrapper[4828]: I1210 20:06:15.765840 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-utilities\") pod \"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa\" (UID: \"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa\") " Dec 10 20:06:15 crc kubenswrapper[4828]: I1210 20:06:15.766453 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-catalog-content\") pod \"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa\" (UID: \"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa\") " Dec 10 20:06:15 crc kubenswrapper[4828]: I1210 20:06:15.766691 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ghvc\" (UniqueName: \"kubernetes.io/projected/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-kube-api-access-5ghvc\") pod \"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa\" (UID: \"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa\") " Dec 10 20:06:15 crc kubenswrapper[4828]: I1210 20:06:15.766943 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-utilities" (OuterVolumeSpecName: "utilities") pod "8fe548f7-e9f5-4bc1-8af3-530d7c5819fa" (UID: "8fe548f7-e9f5-4bc1-8af3-530d7c5819fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:06:15 crc kubenswrapper[4828]: I1210 20:06:15.767608 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:06:15 crc kubenswrapper[4828]: I1210 20:06:15.774463 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-kube-api-access-5ghvc" (OuterVolumeSpecName: "kube-api-access-5ghvc") pod "8fe548f7-e9f5-4bc1-8af3-530d7c5819fa" (UID: "8fe548f7-e9f5-4bc1-8af3-530d7c5819fa"). InnerVolumeSpecName "kube-api-access-5ghvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:06:15 crc kubenswrapper[4828]: I1210 20:06:15.790060 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8fe548f7-e9f5-4bc1-8af3-530d7c5819fa" (UID: "8fe548f7-e9f5-4bc1-8af3-530d7c5819fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:06:15 crc kubenswrapper[4828]: I1210 20:06:15.869518 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ghvc\" (UniqueName: \"kubernetes.io/projected/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-kube-api-access-5ghvc\") on node \"crc\" DevicePath \"\"" Dec 10 20:06:15 crc kubenswrapper[4828]: I1210 20:06:15.869553 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:06:16 crc kubenswrapper[4828]: I1210 20:06:16.188140 4828 generic.go:334] "Generic (PLEG): container finished" podID="8fe548f7-e9f5-4bc1-8af3-530d7c5819fa" containerID="dc8cd3eda11c220972236f9c1dd8e00f8e558f49b3468168d0f6b8c955a9f81e" exitCode=0 Dec 10 20:06:16 crc kubenswrapper[4828]: I1210 20:06:16.188194 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzvrf" event={"ID":"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa","Type":"ContainerDied","Data":"dc8cd3eda11c220972236f9c1dd8e00f8e558f49b3468168d0f6b8c955a9f81e"} Dec 10 20:06:16 crc kubenswrapper[4828]: I1210 20:06:16.188221 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kzvrf" Dec 10 20:06:16 crc kubenswrapper[4828]: I1210 20:06:16.188239 4828 scope.go:117] "RemoveContainer" containerID="dc8cd3eda11c220972236f9c1dd8e00f8e558f49b3468168d0f6b8c955a9f81e" Dec 10 20:06:16 crc kubenswrapper[4828]: I1210 20:06:16.188226 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kzvrf" event={"ID":"8fe548f7-e9f5-4bc1-8af3-530d7c5819fa","Type":"ContainerDied","Data":"38dee5e51b136f452218a107783340667824018010f45de93f861dad3c2b37a0"} Dec 10 20:06:16 crc kubenswrapper[4828]: I1210 20:06:16.216174 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kzvrf"] Dec 10 20:06:16 crc kubenswrapper[4828]: I1210 20:06:16.224588 4828 scope.go:117] "RemoveContainer" containerID="40c834d7c379fd9f8bdf5ea05e1e19e7e4bcd9c2972921a2b27342e1e5f31095" Dec 10 20:06:16 crc kubenswrapper[4828]: I1210 20:06:16.228057 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kzvrf"] Dec 10 20:06:16 crc kubenswrapper[4828]: I1210 20:06:16.248954 4828 scope.go:117] "RemoveContainer" containerID="10c0cd6d8a3a8db050aa2d7fb26e982780dece019dee3c15551a9822832266f4" Dec 10 20:06:16 crc kubenswrapper[4828]: I1210 20:06:16.302646 4828 scope.go:117] "RemoveContainer" containerID="dc8cd3eda11c220972236f9c1dd8e00f8e558f49b3468168d0f6b8c955a9f81e" Dec 10 20:06:16 crc kubenswrapper[4828]: E1210 20:06:16.303038 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc8cd3eda11c220972236f9c1dd8e00f8e558f49b3468168d0f6b8c955a9f81e\": container with ID starting with dc8cd3eda11c220972236f9c1dd8e00f8e558f49b3468168d0f6b8c955a9f81e not found: ID does not exist" containerID="dc8cd3eda11c220972236f9c1dd8e00f8e558f49b3468168d0f6b8c955a9f81e" Dec 10 20:06:16 crc kubenswrapper[4828]: I1210 20:06:16.303065 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc8cd3eda11c220972236f9c1dd8e00f8e558f49b3468168d0f6b8c955a9f81e"} err="failed to get container status \"dc8cd3eda11c220972236f9c1dd8e00f8e558f49b3468168d0f6b8c955a9f81e\": rpc error: code = NotFound desc = could not find container \"dc8cd3eda11c220972236f9c1dd8e00f8e558f49b3468168d0f6b8c955a9f81e\": container with ID starting with dc8cd3eda11c220972236f9c1dd8e00f8e558f49b3468168d0f6b8c955a9f81e not found: ID does not exist" Dec 10 20:06:16 crc kubenswrapper[4828]: I1210 20:06:16.303084 4828 scope.go:117] "RemoveContainer" containerID="40c834d7c379fd9f8bdf5ea05e1e19e7e4bcd9c2972921a2b27342e1e5f31095" Dec 10 20:06:16 crc kubenswrapper[4828]: E1210 20:06:16.303310 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40c834d7c379fd9f8bdf5ea05e1e19e7e4bcd9c2972921a2b27342e1e5f31095\": container with ID starting with 40c834d7c379fd9f8bdf5ea05e1e19e7e4bcd9c2972921a2b27342e1e5f31095 not found: ID does not exist" containerID="40c834d7c379fd9f8bdf5ea05e1e19e7e4bcd9c2972921a2b27342e1e5f31095" Dec 10 20:06:16 crc kubenswrapper[4828]: I1210 20:06:16.303340 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40c834d7c379fd9f8bdf5ea05e1e19e7e4bcd9c2972921a2b27342e1e5f31095"} err="failed to get container status \"40c834d7c379fd9f8bdf5ea05e1e19e7e4bcd9c2972921a2b27342e1e5f31095\": rpc error: code = NotFound desc = could not find container \"40c834d7c379fd9f8bdf5ea05e1e19e7e4bcd9c2972921a2b27342e1e5f31095\": container with ID starting with 40c834d7c379fd9f8bdf5ea05e1e19e7e4bcd9c2972921a2b27342e1e5f31095 not found: ID does not exist" Dec 10 20:06:16 crc kubenswrapper[4828]: I1210 20:06:16.303376 4828 scope.go:117] "RemoveContainer" containerID="10c0cd6d8a3a8db050aa2d7fb26e982780dece019dee3c15551a9822832266f4" Dec 10 20:06:16 crc kubenswrapper[4828]: E1210 20:06:16.303629 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10c0cd6d8a3a8db050aa2d7fb26e982780dece019dee3c15551a9822832266f4\": container with ID starting with 10c0cd6d8a3a8db050aa2d7fb26e982780dece019dee3c15551a9822832266f4 not found: ID does not exist" containerID="10c0cd6d8a3a8db050aa2d7fb26e982780dece019dee3c15551a9822832266f4" Dec 10 20:06:16 crc kubenswrapper[4828]: I1210 20:06:16.303653 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10c0cd6d8a3a8db050aa2d7fb26e982780dece019dee3c15551a9822832266f4"} err="failed to get container status \"10c0cd6d8a3a8db050aa2d7fb26e982780dece019dee3c15551a9822832266f4\": rpc error: code = NotFound desc = could not find container \"10c0cd6d8a3a8db050aa2d7fb26e982780dece019dee3c15551a9822832266f4\": container with ID starting with 10c0cd6d8a3a8db050aa2d7fb26e982780dece019dee3c15551a9822832266f4 not found: ID does not exist" Dec 10 20:06:17 crc kubenswrapper[4828]: I1210 20:06:17.803872 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fe548f7-e9f5-4bc1-8af3-530d7c5819fa" path="/var/lib/kubelet/pods/8fe548f7-e9f5-4bc1-8af3-530d7c5819fa/volumes" Dec 10 20:06:21 crc kubenswrapper[4828]: I1210 20:06:21.230325 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:06:21 crc kubenswrapper[4828]: I1210 20:06:21.231139 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:06:51 crc kubenswrapper[4828]: I1210 20:06:51.230240 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:06:51 crc kubenswrapper[4828]: I1210 20:06:51.230677 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:06:51 crc kubenswrapper[4828]: I1210 20:06:51.230726 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 20:06:51 crc kubenswrapper[4828]: I1210 20:06:51.231644 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 20:06:51 crc kubenswrapper[4828]: I1210 20:06:51.231708 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" gracePeriod=600 Dec 10 20:06:51 crc kubenswrapper[4828]: E1210 20:06:51.367830 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:06:51 crc kubenswrapper[4828]: I1210 20:06:51.587369 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" exitCode=0 Dec 10 20:06:51 crc kubenswrapper[4828]: I1210 20:06:51.587470 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782"} Dec 10 20:06:51 crc kubenswrapper[4828]: I1210 20:06:51.587513 4828 scope.go:117] "RemoveContainer" containerID="db5d83b0d9b828f200e2cc4f0121342bbc11c64408cff1c60141175d4820ef05" Dec 10 20:06:51 crc kubenswrapper[4828]: I1210 20:06:51.588323 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:06:51 crc kubenswrapper[4828]: E1210 20:06:51.588729 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:06:54 crc kubenswrapper[4828]: I1210 20:06:54.894923 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mdhf7"] Dec 10 20:06:54 crc kubenswrapper[4828]: E1210 20:06:54.896205 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe548f7-e9f5-4bc1-8af3-530d7c5819fa" containerName="extract-utilities" Dec 10 20:06:54 crc kubenswrapper[4828]: I1210 20:06:54.896225 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe548f7-e9f5-4bc1-8af3-530d7c5819fa" containerName="extract-utilities" Dec 10 20:06:54 crc kubenswrapper[4828]: E1210 20:06:54.896262 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe548f7-e9f5-4bc1-8af3-530d7c5819fa" containerName="extract-content" Dec 10 20:06:54 crc kubenswrapper[4828]: I1210 20:06:54.896269 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe548f7-e9f5-4bc1-8af3-530d7c5819fa" containerName="extract-content" Dec 10 20:06:54 crc kubenswrapper[4828]: E1210 20:06:54.896316 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe548f7-e9f5-4bc1-8af3-530d7c5819fa" containerName="registry-server" Dec 10 20:06:54 crc kubenswrapper[4828]: I1210 20:06:54.896323 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe548f7-e9f5-4bc1-8af3-530d7c5819fa" containerName="registry-server" Dec 10 20:06:54 crc kubenswrapper[4828]: I1210 20:06:54.896739 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fe548f7-e9f5-4bc1-8af3-530d7c5819fa" containerName="registry-server" Dec 10 20:06:54 crc kubenswrapper[4828]: I1210 20:06:54.898696 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mdhf7" Dec 10 20:06:54 crc kubenswrapper[4828]: I1210 20:06:54.924303 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mdhf7"] Dec 10 20:06:55 crc kubenswrapper[4828]: I1210 20:06:55.023980 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/224e0db3-ed73-4253-9a0c-21564b303e48-utilities\") pod \"community-operators-mdhf7\" (UID: \"224e0db3-ed73-4253-9a0c-21564b303e48\") " pod="openshift-marketplace/community-operators-mdhf7" Dec 10 20:06:55 crc kubenswrapper[4828]: I1210 20:06:55.024055 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/224e0db3-ed73-4253-9a0c-21564b303e48-catalog-content\") pod \"community-operators-mdhf7\" (UID: \"224e0db3-ed73-4253-9a0c-21564b303e48\") " pod="openshift-marketplace/community-operators-mdhf7" Dec 10 20:06:55 crc kubenswrapper[4828]: I1210 20:06:55.024138 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjfx8\" (UniqueName: \"kubernetes.io/projected/224e0db3-ed73-4253-9a0c-21564b303e48-kube-api-access-zjfx8\") pod \"community-operators-mdhf7\" (UID: \"224e0db3-ed73-4253-9a0c-21564b303e48\") " pod="openshift-marketplace/community-operators-mdhf7" Dec 10 20:06:55 crc kubenswrapper[4828]: I1210 20:06:55.126473 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/224e0db3-ed73-4253-9a0c-21564b303e48-utilities\") pod \"community-operators-mdhf7\" (UID: \"224e0db3-ed73-4253-9a0c-21564b303e48\") " pod="openshift-marketplace/community-operators-mdhf7" Dec 10 20:06:55 crc kubenswrapper[4828]: I1210 20:06:55.126572 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/224e0db3-ed73-4253-9a0c-21564b303e48-catalog-content\") pod \"community-operators-mdhf7\" (UID: \"224e0db3-ed73-4253-9a0c-21564b303e48\") " pod="openshift-marketplace/community-operators-mdhf7" Dec 10 20:06:55 crc kubenswrapper[4828]: I1210 20:06:55.126700 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjfx8\" (UniqueName: \"kubernetes.io/projected/224e0db3-ed73-4253-9a0c-21564b303e48-kube-api-access-zjfx8\") pod \"community-operators-mdhf7\" (UID: \"224e0db3-ed73-4253-9a0c-21564b303e48\") " pod="openshift-marketplace/community-operators-mdhf7" Dec 10 20:06:55 crc kubenswrapper[4828]: I1210 20:06:55.127192 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/224e0db3-ed73-4253-9a0c-21564b303e48-utilities\") pod \"community-operators-mdhf7\" (UID: \"224e0db3-ed73-4253-9a0c-21564b303e48\") " pod="openshift-marketplace/community-operators-mdhf7" Dec 10 20:06:55 crc kubenswrapper[4828]: I1210 20:06:55.127329 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/224e0db3-ed73-4253-9a0c-21564b303e48-catalog-content\") pod \"community-operators-mdhf7\" (UID: \"224e0db3-ed73-4253-9a0c-21564b303e48\") " pod="openshift-marketplace/community-operators-mdhf7" Dec 10 20:06:55 crc kubenswrapper[4828]: I1210 20:06:55.155091 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjfx8\" (UniqueName: \"kubernetes.io/projected/224e0db3-ed73-4253-9a0c-21564b303e48-kube-api-access-zjfx8\") pod \"community-operators-mdhf7\" (UID: \"224e0db3-ed73-4253-9a0c-21564b303e48\") " pod="openshift-marketplace/community-operators-mdhf7" Dec 10 20:06:55 crc kubenswrapper[4828]: I1210 20:06:55.221734 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mdhf7" Dec 10 20:06:55 crc kubenswrapper[4828]: W1210 20:06:55.759729 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod224e0db3_ed73_4253_9a0c_21564b303e48.slice/crio-b2bf54545868102ad1d0c0edf4e9993c5614c1f8a96a4c9561f361bf14d165ec WatchSource:0}: Error finding container b2bf54545868102ad1d0c0edf4e9993c5614c1f8a96a4c9561f361bf14d165ec: Status 404 returned error can't find the container with id b2bf54545868102ad1d0c0edf4e9993c5614c1f8a96a4c9561f361bf14d165ec Dec 10 20:06:55 crc kubenswrapper[4828]: I1210 20:06:55.764392 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mdhf7"] Dec 10 20:06:56 crc kubenswrapper[4828]: I1210 20:06:56.641479 4828 generic.go:334] "Generic (PLEG): container finished" podID="224e0db3-ed73-4253-9a0c-21564b303e48" containerID="4460525a274ee3642a7f084dd31a6bcc4a0d7e4d1ad12bc66eee018ae1bfdebf" exitCode=0 Dec 10 20:06:56 crc kubenswrapper[4828]: I1210 20:06:56.641539 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mdhf7" event={"ID":"224e0db3-ed73-4253-9a0c-21564b303e48","Type":"ContainerDied","Data":"4460525a274ee3642a7f084dd31a6bcc4a0d7e4d1ad12bc66eee018ae1bfdebf"} Dec 10 20:06:56 crc kubenswrapper[4828]: I1210 20:06:56.641982 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mdhf7" event={"ID":"224e0db3-ed73-4253-9a0c-21564b303e48","Type":"ContainerStarted","Data":"b2bf54545868102ad1d0c0edf4e9993c5614c1f8a96a4c9561f361bf14d165ec"} Dec 10 20:06:58 crc kubenswrapper[4828]: I1210 20:06:58.669273 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mdhf7" event={"ID":"224e0db3-ed73-4253-9a0c-21564b303e48","Type":"ContainerStarted","Data":"8629c99be00623ceca51946db63841e1970143da3a8da4767211141b1ef98e0c"} Dec 10 20:06:59 crc kubenswrapper[4828]: I1210 20:06:59.680275 4828 generic.go:334] "Generic (PLEG): container finished" podID="224e0db3-ed73-4253-9a0c-21564b303e48" containerID="8629c99be00623ceca51946db63841e1970143da3a8da4767211141b1ef98e0c" exitCode=0 Dec 10 20:06:59 crc kubenswrapper[4828]: I1210 20:06:59.680326 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mdhf7" event={"ID":"224e0db3-ed73-4253-9a0c-21564b303e48","Type":"ContainerDied","Data":"8629c99be00623ceca51946db63841e1970143da3a8da4767211141b1ef98e0c"} Dec 10 20:07:01 crc kubenswrapper[4828]: I1210 20:07:01.701612 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mdhf7" event={"ID":"224e0db3-ed73-4253-9a0c-21564b303e48","Type":"ContainerStarted","Data":"0ee04de3728c74889f63b6492feda797889ef5ba58bba086bd05f314fe77090d"} Dec 10 20:07:01 crc kubenswrapper[4828]: I1210 20:07:01.726884 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mdhf7" podStartSLOduration=4.104540395 podStartE2EDuration="7.726866856s" podCreationTimestamp="2025-12-10 20:06:54 +0000 UTC" firstStartedPulling="2025-12-10 20:06:56.644346555 +0000 UTC m=+4297.154957560" lastFinishedPulling="2025-12-10 20:07:00.266673016 +0000 UTC m=+4300.777284021" observedRunningTime="2025-12-10 20:07:01.716390377 +0000 UTC m=+4302.227001432" watchObservedRunningTime="2025-12-10 20:07:01.726866856 +0000 UTC m=+4302.237477861" Dec 10 20:07:03 crc kubenswrapper[4828]: I1210 20:07:03.788960 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:07:03 crc kubenswrapper[4828]: E1210 20:07:03.789922 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:07:05 crc kubenswrapper[4828]: I1210 20:07:05.222567 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mdhf7" Dec 10 20:07:05 crc kubenswrapper[4828]: I1210 20:07:05.222875 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mdhf7" Dec 10 20:07:05 crc kubenswrapper[4828]: I1210 20:07:05.269307 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mdhf7" Dec 10 20:07:05 crc kubenswrapper[4828]: I1210 20:07:05.812863 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mdhf7" Dec 10 20:07:05 crc kubenswrapper[4828]: I1210 20:07:05.855753 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mdhf7"] Dec 10 20:07:07 crc kubenswrapper[4828]: I1210 20:07:07.787684 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mdhf7" podUID="224e0db3-ed73-4253-9a0c-21564b303e48" containerName="registry-server" containerID="cri-o://0ee04de3728c74889f63b6492feda797889ef5ba58bba086bd05f314fe77090d" gracePeriod=2 Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.336514 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mdhf7" Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.471328 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjfx8\" (UniqueName: \"kubernetes.io/projected/224e0db3-ed73-4253-9a0c-21564b303e48-kube-api-access-zjfx8\") pod \"224e0db3-ed73-4253-9a0c-21564b303e48\" (UID: \"224e0db3-ed73-4253-9a0c-21564b303e48\") " Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.471447 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/224e0db3-ed73-4253-9a0c-21564b303e48-utilities\") pod \"224e0db3-ed73-4253-9a0c-21564b303e48\" (UID: \"224e0db3-ed73-4253-9a0c-21564b303e48\") " Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.471739 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/224e0db3-ed73-4253-9a0c-21564b303e48-catalog-content\") pod \"224e0db3-ed73-4253-9a0c-21564b303e48\" (UID: \"224e0db3-ed73-4253-9a0c-21564b303e48\") " Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.473516 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/224e0db3-ed73-4253-9a0c-21564b303e48-utilities" (OuterVolumeSpecName: "utilities") pod "224e0db3-ed73-4253-9a0c-21564b303e48" (UID: "224e0db3-ed73-4253-9a0c-21564b303e48"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.474594 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/224e0db3-ed73-4253-9a0c-21564b303e48-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.478322 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/224e0db3-ed73-4253-9a0c-21564b303e48-kube-api-access-zjfx8" (OuterVolumeSpecName: "kube-api-access-zjfx8") pod "224e0db3-ed73-4253-9a0c-21564b303e48" (UID: "224e0db3-ed73-4253-9a0c-21564b303e48"). InnerVolumeSpecName "kube-api-access-zjfx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.517324 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/224e0db3-ed73-4253-9a0c-21564b303e48-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "224e0db3-ed73-4253-9a0c-21564b303e48" (UID: "224e0db3-ed73-4253-9a0c-21564b303e48"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.577141 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/224e0db3-ed73-4253-9a0c-21564b303e48-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.577468 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjfx8\" (UniqueName: \"kubernetes.io/projected/224e0db3-ed73-4253-9a0c-21564b303e48-kube-api-access-zjfx8\") on node \"crc\" DevicePath \"\"" Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.800998 4828 generic.go:334] "Generic (PLEG): container finished" podID="224e0db3-ed73-4253-9a0c-21564b303e48" containerID="0ee04de3728c74889f63b6492feda797889ef5ba58bba086bd05f314fe77090d" exitCode=0 Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.801046 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mdhf7" event={"ID":"224e0db3-ed73-4253-9a0c-21564b303e48","Type":"ContainerDied","Data":"0ee04de3728c74889f63b6492feda797889ef5ba58bba086bd05f314fe77090d"} Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.801066 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mdhf7" Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.801092 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mdhf7" event={"ID":"224e0db3-ed73-4253-9a0c-21564b303e48","Type":"ContainerDied","Data":"b2bf54545868102ad1d0c0edf4e9993c5614c1f8a96a4c9561f361bf14d165ec"} Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.801110 4828 scope.go:117] "RemoveContainer" containerID="0ee04de3728c74889f63b6492feda797889ef5ba58bba086bd05f314fe77090d" Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.822544 4828 scope.go:117] "RemoveContainer" containerID="8629c99be00623ceca51946db63841e1970143da3a8da4767211141b1ef98e0c" Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.847295 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mdhf7"] Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.860127 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mdhf7"] Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.862230 4828 scope.go:117] "RemoveContainer" containerID="4460525a274ee3642a7f084dd31a6bcc4a0d7e4d1ad12bc66eee018ae1bfdebf" Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.905285 4828 scope.go:117] "RemoveContainer" containerID="0ee04de3728c74889f63b6492feda797889ef5ba58bba086bd05f314fe77090d" Dec 10 20:07:08 crc kubenswrapper[4828]: E1210 20:07:08.905698 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ee04de3728c74889f63b6492feda797889ef5ba58bba086bd05f314fe77090d\": container with ID starting with 0ee04de3728c74889f63b6492feda797889ef5ba58bba086bd05f314fe77090d not found: ID does not exist" containerID="0ee04de3728c74889f63b6492feda797889ef5ba58bba086bd05f314fe77090d" Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.905730 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ee04de3728c74889f63b6492feda797889ef5ba58bba086bd05f314fe77090d"} err="failed to get container status \"0ee04de3728c74889f63b6492feda797889ef5ba58bba086bd05f314fe77090d\": rpc error: code = NotFound desc = could not find container \"0ee04de3728c74889f63b6492feda797889ef5ba58bba086bd05f314fe77090d\": container with ID starting with 0ee04de3728c74889f63b6492feda797889ef5ba58bba086bd05f314fe77090d not found: ID does not exist" Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.905752 4828 scope.go:117] "RemoveContainer" containerID="8629c99be00623ceca51946db63841e1970143da3a8da4767211141b1ef98e0c" Dec 10 20:07:08 crc kubenswrapper[4828]: E1210 20:07:08.906030 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8629c99be00623ceca51946db63841e1970143da3a8da4767211141b1ef98e0c\": container with ID starting with 8629c99be00623ceca51946db63841e1970143da3a8da4767211141b1ef98e0c not found: ID does not exist" containerID="8629c99be00623ceca51946db63841e1970143da3a8da4767211141b1ef98e0c" Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.906051 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8629c99be00623ceca51946db63841e1970143da3a8da4767211141b1ef98e0c"} err="failed to get container status \"8629c99be00623ceca51946db63841e1970143da3a8da4767211141b1ef98e0c\": rpc error: code = NotFound desc = could not find container \"8629c99be00623ceca51946db63841e1970143da3a8da4767211141b1ef98e0c\": container with ID starting with 8629c99be00623ceca51946db63841e1970143da3a8da4767211141b1ef98e0c not found: ID does not exist" Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.906065 4828 scope.go:117] "RemoveContainer" containerID="4460525a274ee3642a7f084dd31a6bcc4a0d7e4d1ad12bc66eee018ae1bfdebf" Dec 10 20:07:08 crc kubenswrapper[4828]: E1210 20:07:08.906419 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4460525a274ee3642a7f084dd31a6bcc4a0d7e4d1ad12bc66eee018ae1bfdebf\": container with ID starting with 4460525a274ee3642a7f084dd31a6bcc4a0d7e4d1ad12bc66eee018ae1bfdebf not found: ID does not exist" containerID="4460525a274ee3642a7f084dd31a6bcc4a0d7e4d1ad12bc66eee018ae1bfdebf" Dec 10 20:07:08 crc kubenswrapper[4828]: I1210 20:07:08.906439 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4460525a274ee3642a7f084dd31a6bcc4a0d7e4d1ad12bc66eee018ae1bfdebf"} err="failed to get container status \"4460525a274ee3642a7f084dd31a6bcc4a0d7e4d1ad12bc66eee018ae1bfdebf\": rpc error: code = NotFound desc = could not find container \"4460525a274ee3642a7f084dd31a6bcc4a0d7e4d1ad12bc66eee018ae1bfdebf\": container with ID starting with 4460525a274ee3642a7f084dd31a6bcc4a0d7e4d1ad12bc66eee018ae1bfdebf not found: ID does not exist" Dec 10 20:07:09 crc kubenswrapper[4828]: I1210 20:07:09.802974 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="224e0db3-ed73-4253-9a0c-21564b303e48" path="/var/lib/kubelet/pods/224e0db3-ed73-4253-9a0c-21564b303e48/volumes" Dec 10 20:07:16 crc kubenswrapper[4828]: I1210 20:07:16.789602 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:07:16 crc kubenswrapper[4828]: E1210 20:07:16.790364 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:07:27 crc kubenswrapper[4828]: I1210 20:07:27.789404 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:07:27 crc kubenswrapper[4828]: E1210 20:07:27.790234 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:07:39 crc kubenswrapper[4828]: I1210 20:07:39.796147 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:07:39 crc kubenswrapper[4828]: E1210 20:07:39.797156 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:07:52 crc kubenswrapper[4828]: I1210 20:07:52.790071 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:07:52 crc kubenswrapper[4828]: E1210 20:07:52.790845 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:08:04 crc kubenswrapper[4828]: I1210 20:08:04.790028 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:08:04 crc kubenswrapper[4828]: E1210 20:08:04.793120 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:08:19 crc kubenswrapper[4828]: I1210 20:08:19.797034 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:08:19 crc kubenswrapper[4828]: E1210 20:08:19.800332 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:08:34 crc kubenswrapper[4828]: I1210 20:08:34.791401 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:08:34 crc kubenswrapper[4828]: E1210 20:08:34.793365 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:08:46 crc kubenswrapper[4828]: I1210 20:08:46.789433 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:08:46 crc kubenswrapper[4828]: E1210 20:08:46.790256 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:09:00 crc kubenswrapper[4828]: I1210 20:09:00.791826 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:09:00 crc kubenswrapper[4828]: E1210 20:09:00.792723 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:09:14 crc kubenswrapper[4828]: I1210 20:09:14.789141 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:09:14 crc kubenswrapper[4828]: E1210 20:09:14.790004 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:09:27 crc kubenswrapper[4828]: I1210 20:09:27.789685 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:09:27 crc kubenswrapper[4828]: E1210 20:09:27.790495 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:09:42 crc kubenswrapper[4828]: I1210 20:09:42.788788 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:09:42 crc kubenswrapper[4828]: E1210 20:09:42.789897 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:09:53 crc kubenswrapper[4828]: I1210 20:09:53.789422 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:09:53 crc kubenswrapper[4828]: E1210 20:09:53.790132 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:10:05 crc kubenswrapper[4828]: I1210 20:10:05.788428 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:10:05 crc kubenswrapper[4828]: E1210 20:10:05.789257 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:10:19 crc kubenswrapper[4828]: I1210 20:10:19.797823 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:10:19 crc kubenswrapper[4828]: E1210 20:10:19.798848 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:10:31 crc kubenswrapper[4828]: I1210 20:10:31.789241 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:10:31 crc kubenswrapper[4828]: E1210 20:10:31.790142 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:10:43 crc kubenswrapper[4828]: I1210 20:10:43.789420 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:10:43 crc kubenswrapper[4828]: E1210 20:10:43.790371 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:10:55 crc kubenswrapper[4828]: I1210 20:10:55.788956 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:10:55 crc kubenswrapper[4828]: E1210 20:10:55.789739 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:11:02 crc kubenswrapper[4828]: I1210 20:11:02.921187 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 10 20:11:02 crc kubenswrapper[4828]: E1210 20:11:02.922345 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="224e0db3-ed73-4253-9a0c-21564b303e48" containerName="extract-utilities" Dec 10 20:11:02 crc kubenswrapper[4828]: I1210 20:11:02.922363 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="224e0db3-ed73-4253-9a0c-21564b303e48" containerName="extract-utilities" Dec 10 20:11:02 crc kubenswrapper[4828]: E1210 20:11:02.922438 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="224e0db3-ed73-4253-9a0c-21564b303e48" containerName="registry-server" Dec 10 20:11:02 crc kubenswrapper[4828]: I1210 20:11:02.922447 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="224e0db3-ed73-4253-9a0c-21564b303e48" containerName="registry-server" Dec 10 20:11:02 crc kubenswrapper[4828]: E1210 20:11:02.922467 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="224e0db3-ed73-4253-9a0c-21564b303e48" containerName="extract-content" Dec 10 20:11:02 crc kubenswrapper[4828]: I1210 20:11:02.922476 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="224e0db3-ed73-4253-9a0c-21564b303e48" containerName="extract-content" Dec 10 20:11:02 crc kubenswrapper[4828]: I1210 20:11:02.922819 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="224e0db3-ed73-4253-9a0c-21564b303e48" containerName="registry-server" Dec 10 20:11:02 crc kubenswrapper[4828]: I1210 20:11:02.923788 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 20:11:02 crc kubenswrapper[4828]: I1210 20:11:02.927856 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 10 20:11:02 crc kubenswrapper[4828]: I1210 20:11:02.928898 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 10 20:11:02 crc kubenswrapper[4828]: I1210 20:11:02.929677 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-zghwg" Dec 10 20:11:02 crc kubenswrapper[4828]: I1210 20:11:02.932178 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 10 20:11:02 crc kubenswrapper[4828]: I1210 20:11:02.933504 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.082851 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/169c0824-0503-4c49-9980-7d045db64b8b-config-data\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.082901 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.082931 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.083032 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.083060 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/169c0824-0503-4c49-9980-7d045db64b8b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.083109 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/169c0824-0503-4c49-9980-7d045db64b8b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.083128 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.083350 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/169c0824-0503-4c49-9980-7d045db64b8b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.083598 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7hvt\" (UniqueName: \"kubernetes.io/projected/169c0824-0503-4c49-9980-7d045db64b8b-kube-api-access-g7hvt\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.186134 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7hvt\" (UniqueName: \"kubernetes.io/projected/169c0824-0503-4c49-9980-7d045db64b8b-kube-api-access-g7hvt\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.186255 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/169c0824-0503-4c49-9980-7d045db64b8b-config-data\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.186285 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.186312 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.186349 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.186373 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/169c0824-0503-4c49-9980-7d045db64b8b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.186420 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/169c0824-0503-4c49-9980-7d045db64b8b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.186437 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.186480 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/169c0824-0503-4c49-9980-7d045db64b8b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.186991 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/169c0824-0503-4c49-9980-7d045db64b8b-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.187196 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/169c0824-0503-4c49-9980-7d045db64b8b-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.189531 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/169c0824-0503-4c49-9980-7d045db64b8b-config-data\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.190100 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/169c0824-0503-4c49-9980-7d045db64b8b-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.191329 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.874723 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.874763 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.874909 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:03 crc kubenswrapper[4828]: I1210 20:11:03.875473 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7hvt\" (UniqueName: \"kubernetes.io/projected/169c0824-0503-4c49-9980-7d045db64b8b-kube-api-access-g7hvt\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:04 crc kubenswrapper[4828]: I1210 20:11:04.105538 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " pod="openstack/tempest-tests-tempest" Dec 10 20:11:04 crc kubenswrapper[4828]: I1210 20:11:04.156661 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 20:11:04 crc kubenswrapper[4828]: I1210 20:11:04.636242 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 10 20:11:04 crc kubenswrapper[4828]: I1210 20:11:04.642457 4828 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 20:11:05 crc kubenswrapper[4828]: I1210 20:11:05.439847 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"169c0824-0503-4c49-9980-7d045db64b8b","Type":"ContainerStarted","Data":"741f6fa6ea23b23dccef289a99c626be723d7212950ad01faf1eb6a631d116ad"} Dec 10 20:11:08 crc kubenswrapper[4828]: I1210 20:11:08.789630 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:11:08 crc kubenswrapper[4828]: E1210 20:11:08.790637 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:11:21 crc kubenswrapper[4828]: I1210 20:11:21.789402 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:11:21 crc kubenswrapper[4828]: E1210 20:11:21.790499 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:11:34 crc kubenswrapper[4828]: E1210 20:11:34.828184 4828 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 10 20:11:34 crc kubenswrapper[4828]: E1210 20:11:34.829662 4828 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g7hvt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(169c0824-0503-4c49-9980-7d045db64b8b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 10 20:11:34 crc kubenswrapper[4828]: E1210 20:11:34.830907 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="169c0824-0503-4c49-9980-7d045db64b8b" Dec 10 20:11:35 crc kubenswrapper[4828]: I1210 20:11:35.789900 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:11:35 crc kubenswrapper[4828]: E1210 20:11:35.790491 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:11:35 crc kubenswrapper[4828]: E1210 20:11:35.793316 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="169c0824-0503-4c49-9980-7d045db64b8b" Dec 10 20:11:47 crc kubenswrapper[4828]: I1210 20:11:47.413197 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 10 20:11:47 crc kubenswrapper[4828]: I1210 20:11:47.663275 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xzmsc"] Dec 10 20:11:47 crc kubenswrapper[4828]: I1210 20:11:47.665746 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzmsc" Dec 10 20:11:47 crc kubenswrapper[4828]: I1210 20:11:47.695018 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xzmsc"] Dec 10 20:11:47 crc kubenswrapper[4828]: I1210 20:11:47.824496 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c93e426a-10a1-46cd-a526-134d4f3d603d-catalog-content\") pod \"certified-operators-xzmsc\" (UID: \"c93e426a-10a1-46cd-a526-134d4f3d603d\") " pod="openshift-marketplace/certified-operators-xzmsc" Dec 10 20:11:47 crc kubenswrapper[4828]: I1210 20:11:47.824549 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c93e426a-10a1-46cd-a526-134d4f3d603d-utilities\") pod \"certified-operators-xzmsc\" (UID: \"c93e426a-10a1-46cd-a526-134d4f3d603d\") " pod="openshift-marketplace/certified-operators-xzmsc" Dec 10 20:11:47 crc kubenswrapper[4828]: I1210 20:11:47.824576 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5vqd\" (UniqueName: \"kubernetes.io/projected/c93e426a-10a1-46cd-a526-134d4f3d603d-kube-api-access-b5vqd\") pod \"certified-operators-xzmsc\" (UID: \"c93e426a-10a1-46cd-a526-134d4f3d603d\") " pod="openshift-marketplace/certified-operators-xzmsc" Dec 10 20:11:47 crc kubenswrapper[4828]: I1210 20:11:47.927788 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c93e426a-10a1-46cd-a526-134d4f3d603d-catalog-content\") pod \"certified-operators-xzmsc\" (UID: \"c93e426a-10a1-46cd-a526-134d4f3d603d\") " pod="openshift-marketplace/certified-operators-xzmsc" Dec 10 20:11:47 crc kubenswrapper[4828]: I1210 20:11:47.928372 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c93e426a-10a1-46cd-a526-134d4f3d603d-utilities\") pod \"certified-operators-xzmsc\" (UID: \"c93e426a-10a1-46cd-a526-134d4f3d603d\") " pod="openshift-marketplace/certified-operators-xzmsc" Dec 10 20:11:47 crc kubenswrapper[4828]: I1210 20:11:47.928478 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5vqd\" (UniqueName: \"kubernetes.io/projected/c93e426a-10a1-46cd-a526-134d4f3d603d-kube-api-access-b5vqd\") pod \"certified-operators-xzmsc\" (UID: \"c93e426a-10a1-46cd-a526-134d4f3d603d\") " pod="openshift-marketplace/certified-operators-xzmsc" Dec 10 20:11:47 crc kubenswrapper[4828]: I1210 20:11:47.928767 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c93e426a-10a1-46cd-a526-134d4f3d603d-utilities\") pod \"certified-operators-xzmsc\" (UID: \"c93e426a-10a1-46cd-a526-134d4f3d603d\") " pod="openshift-marketplace/certified-operators-xzmsc" Dec 10 20:11:47 crc kubenswrapper[4828]: I1210 20:11:47.928860 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c93e426a-10a1-46cd-a526-134d4f3d603d-catalog-content\") pod \"certified-operators-xzmsc\" (UID: \"c93e426a-10a1-46cd-a526-134d4f3d603d\") " pod="openshift-marketplace/certified-operators-xzmsc" Dec 10 20:11:47 crc kubenswrapper[4828]: I1210 20:11:47.959876 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5vqd\" (UniqueName: \"kubernetes.io/projected/c93e426a-10a1-46cd-a526-134d4f3d603d-kube-api-access-b5vqd\") pod \"certified-operators-xzmsc\" (UID: \"c93e426a-10a1-46cd-a526-134d4f3d603d\") " pod="openshift-marketplace/certified-operators-xzmsc" Dec 10 20:11:48 crc kubenswrapper[4828]: I1210 20:11:48.004908 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzmsc" Dec 10 20:11:48 crc kubenswrapper[4828]: I1210 20:11:48.588332 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xzmsc"] Dec 10 20:11:48 crc kubenswrapper[4828]: I1210 20:11:48.789118 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:11:48 crc kubenswrapper[4828]: E1210 20:11:48.789756 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:11:48 crc kubenswrapper[4828]: I1210 20:11:48.937444 4828 generic.go:334] "Generic (PLEG): container finished" podID="c93e426a-10a1-46cd-a526-134d4f3d603d" containerID="b1f5f22118ecd536fe751b59ebeb5d1d1496a2d095d66cfa5d5a89c865e9ca72" exitCode=0 Dec 10 20:11:48 crc kubenswrapper[4828]: I1210 20:11:48.937497 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzmsc" event={"ID":"c93e426a-10a1-46cd-a526-134d4f3d603d","Type":"ContainerDied","Data":"b1f5f22118ecd536fe751b59ebeb5d1d1496a2d095d66cfa5d5a89c865e9ca72"} Dec 10 20:11:48 crc kubenswrapper[4828]: I1210 20:11:48.937547 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzmsc" event={"ID":"c93e426a-10a1-46cd-a526-134d4f3d603d","Type":"ContainerStarted","Data":"cb0dc32f17a6515c44cfd3a634674c5a660849c6f4ab56d7dd434155949c0b52"} Dec 10 20:11:48 crc kubenswrapper[4828]: I1210 20:11:48.941742 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"169c0824-0503-4c49-9980-7d045db64b8b","Type":"ContainerStarted","Data":"b0ded671bccc9c9ec6244ca39ec5ad6af9608530e8dce8290ac1dc4cf7764f65"} Dec 10 20:11:48 crc kubenswrapper[4828]: I1210 20:11:48.978617 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=5.210477272 podStartE2EDuration="47.978600432s" podCreationTimestamp="2025-12-10 20:11:01 +0000 UTC" firstStartedPulling="2025-12-10 20:11:04.642167435 +0000 UTC m=+4545.152778460" lastFinishedPulling="2025-12-10 20:11:47.410290615 +0000 UTC m=+4587.920901620" observedRunningTime="2025-12-10 20:11:48.972714676 +0000 UTC m=+4589.483325681" watchObservedRunningTime="2025-12-10 20:11:48.978600432 +0000 UTC m=+4589.489211437" Dec 10 20:11:50 crc kubenswrapper[4828]: I1210 20:11:50.967537 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzmsc" event={"ID":"c93e426a-10a1-46cd-a526-134d4f3d603d","Type":"ContainerStarted","Data":"a3e7739cc24a3123cbff00e8ba834c1a2709be57f525ed1c487b5885bb8c1c9b"} Dec 10 20:11:51 crc kubenswrapper[4828]: I1210 20:11:51.980398 4828 generic.go:334] "Generic (PLEG): container finished" podID="c93e426a-10a1-46cd-a526-134d4f3d603d" containerID="a3e7739cc24a3123cbff00e8ba834c1a2709be57f525ed1c487b5885bb8c1c9b" exitCode=0 Dec 10 20:11:51 crc kubenswrapper[4828]: I1210 20:11:51.980442 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzmsc" event={"ID":"c93e426a-10a1-46cd-a526-134d4f3d603d","Type":"ContainerDied","Data":"a3e7739cc24a3123cbff00e8ba834c1a2709be57f525ed1c487b5885bb8c1c9b"} Dec 10 20:11:52 crc kubenswrapper[4828]: I1210 20:11:52.991884 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzmsc" event={"ID":"c93e426a-10a1-46cd-a526-134d4f3d603d","Type":"ContainerStarted","Data":"717ccec5211da0e4f6adffd3c7245f0ab258e72e7afaa78b59861ad68aa82c99"} Dec 10 20:11:53 crc kubenswrapper[4828]: I1210 20:11:53.017070 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xzmsc" podStartSLOduration=2.525371198 podStartE2EDuration="6.017052156s" podCreationTimestamp="2025-12-10 20:11:47 +0000 UTC" firstStartedPulling="2025-12-10 20:11:48.939430041 +0000 UTC m=+4589.450041036" lastFinishedPulling="2025-12-10 20:11:52.431110979 +0000 UTC m=+4592.941721994" observedRunningTime="2025-12-10 20:11:53.007371759 +0000 UTC m=+4593.517982774" watchObservedRunningTime="2025-12-10 20:11:53.017052156 +0000 UTC m=+4593.527663161" Dec 10 20:11:58 crc kubenswrapper[4828]: I1210 20:11:58.006041 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xzmsc" Dec 10 20:11:58 crc kubenswrapper[4828]: I1210 20:11:58.006558 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xzmsc" Dec 10 20:11:58 crc kubenswrapper[4828]: I1210 20:11:58.055016 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xzmsc" Dec 10 20:11:58 crc kubenswrapper[4828]: I1210 20:11:58.110571 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xzmsc" Dec 10 20:11:58 crc kubenswrapper[4828]: I1210 20:11:58.295694 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xzmsc"] Dec 10 20:11:59 crc kubenswrapper[4828]: I1210 20:11:59.799515 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:12:00 crc kubenswrapper[4828]: I1210 20:12:00.067348 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"7918ddadd3ea00c849d997afdc8c99d59b1ca27ab5e5de92d512e606ec4f36b8"} Dec 10 20:12:00 crc kubenswrapper[4828]: I1210 20:12:00.067515 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xzmsc" podUID="c93e426a-10a1-46cd-a526-134d4f3d603d" containerName="registry-server" containerID="cri-o://717ccec5211da0e4f6adffd3c7245f0ab258e72e7afaa78b59861ad68aa82c99" gracePeriod=2 Dec 10 20:12:00 crc kubenswrapper[4828]: I1210 20:12:00.622024 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzmsc" Dec 10 20:12:00 crc kubenswrapper[4828]: I1210 20:12:00.753085 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5vqd\" (UniqueName: \"kubernetes.io/projected/c93e426a-10a1-46cd-a526-134d4f3d603d-kube-api-access-b5vqd\") pod \"c93e426a-10a1-46cd-a526-134d4f3d603d\" (UID: \"c93e426a-10a1-46cd-a526-134d4f3d603d\") " Dec 10 20:12:00 crc kubenswrapper[4828]: I1210 20:12:00.753553 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c93e426a-10a1-46cd-a526-134d4f3d603d-catalog-content\") pod \"c93e426a-10a1-46cd-a526-134d4f3d603d\" (UID: \"c93e426a-10a1-46cd-a526-134d4f3d603d\") " Dec 10 20:12:00 crc kubenswrapper[4828]: I1210 20:12:00.753735 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c93e426a-10a1-46cd-a526-134d4f3d603d-utilities\") pod \"c93e426a-10a1-46cd-a526-134d4f3d603d\" (UID: \"c93e426a-10a1-46cd-a526-134d4f3d603d\") " Dec 10 20:12:00 crc kubenswrapper[4828]: I1210 20:12:00.755181 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c93e426a-10a1-46cd-a526-134d4f3d603d-utilities" (OuterVolumeSpecName: "utilities") pod "c93e426a-10a1-46cd-a526-134d4f3d603d" (UID: "c93e426a-10a1-46cd-a526-134d4f3d603d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:12:00 crc kubenswrapper[4828]: I1210 20:12:00.761189 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c93e426a-10a1-46cd-a526-134d4f3d603d-kube-api-access-b5vqd" (OuterVolumeSpecName: "kube-api-access-b5vqd") pod "c93e426a-10a1-46cd-a526-134d4f3d603d" (UID: "c93e426a-10a1-46cd-a526-134d4f3d603d"). InnerVolumeSpecName "kube-api-access-b5vqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:12:00 crc kubenswrapper[4828]: I1210 20:12:00.820564 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c93e426a-10a1-46cd-a526-134d4f3d603d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c93e426a-10a1-46cd-a526-134d4f3d603d" (UID: "c93e426a-10a1-46cd-a526-134d4f3d603d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:12:00 crc kubenswrapper[4828]: I1210 20:12:00.857095 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c93e426a-10a1-46cd-a526-134d4f3d603d-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:12:00 crc kubenswrapper[4828]: I1210 20:12:00.857135 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5vqd\" (UniqueName: \"kubernetes.io/projected/c93e426a-10a1-46cd-a526-134d4f3d603d-kube-api-access-b5vqd\") on node \"crc\" DevicePath \"\"" Dec 10 20:12:00 crc kubenswrapper[4828]: I1210 20:12:00.857148 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c93e426a-10a1-46cd-a526-134d4f3d603d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:12:01 crc kubenswrapper[4828]: I1210 20:12:01.083862 4828 generic.go:334] "Generic (PLEG): container finished" podID="c93e426a-10a1-46cd-a526-134d4f3d603d" containerID="717ccec5211da0e4f6adffd3c7245f0ab258e72e7afaa78b59861ad68aa82c99" exitCode=0 Dec 10 20:12:01 crc kubenswrapper[4828]: I1210 20:12:01.083905 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzmsc" event={"ID":"c93e426a-10a1-46cd-a526-134d4f3d603d","Type":"ContainerDied","Data":"717ccec5211da0e4f6adffd3c7245f0ab258e72e7afaa78b59861ad68aa82c99"} Dec 10 20:12:01 crc kubenswrapper[4828]: I1210 20:12:01.083923 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzmsc" Dec 10 20:12:01 crc kubenswrapper[4828]: I1210 20:12:01.083944 4828 scope.go:117] "RemoveContainer" containerID="717ccec5211da0e4f6adffd3c7245f0ab258e72e7afaa78b59861ad68aa82c99" Dec 10 20:12:01 crc kubenswrapper[4828]: I1210 20:12:01.083933 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzmsc" event={"ID":"c93e426a-10a1-46cd-a526-134d4f3d603d","Type":"ContainerDied","Data":"cb0dc32f17a6515c44cfd3a634674c5a660849c6f4ab56d7dd434155949c0b52"} Dec 10 20:12:01 crc kubenswrapper[4828]: I1210 20:12:01.123529 4828 scope.go:117] "RemoveContainer" containerID="a3e7739cc24a3123cbff00e8ba834c1a2709be57f525ed1c487b5885bb8c1c9b" Dec 10 20:12:01 crc kubenswrapper[4828]: I1210 20:12:01.137687 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xzmsc"] Dec 10 20:12:01 crc kubenswrapper[4828]: I1210 20:12:01.151220 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xzmsc"] Dec 10 20:12:01 crc kubenswrapper[4828]: I1210 20:12:01.155032 4828 scope.go:117] "RemoveContainer" containerID="b1f5f22118ecd536fe751b59ebeb5d1d1496a2d095d66cfa5d5a89c865e9ca72" Dec 10 20:12:01 crc kubenswrapper[4828]: I1210 20:12:01.213524 4828 scope.go:117] "RemoveContainer" containerID="717ccec5211da0e4f6adffd3c7245f0ab258e72e7afaa78b59861ad68aa82c99" Dec 10 20:12:01 crc kubenswrapper[4828]: E1210 20:12:01.213997 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"717ccec5211da0e4f6adffd3c7245f0ab258e72e7afaa78b59861ad68aa82c99\": container with ID starting with 717ccec5211da0e4f6adffd3c7245f0ab258e72e7afaa78b59861ad68aa82c99 not found: ID does not exist" containerID="717ccec5211da0e4f6adffd3c7245f0ab258e72e7afaa78b59861ad68aa82c99" Dec 10 20:12:01 crc kubenswrapper[4828]: I1210 20:12:01.214051 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"717ccec5211da0e4f6adffd3c7245f0ab258e72e7afaa78b59861ad68aa82c99"} err="failed to get container status \"717ccec5211da0e4f6adffd3c7245f0ab258e72e7afaa78b59861ad68aa82c99\": rpc error: code = NotFound desc = could not find container \"717ccec5211da0e4f6adffd3c7245f0ab258e72e7afaa78b59861ad68aa82c99\": container with ID starting with 717ccec5211da0e4f6adffd3c7245f0ab258e72e7afaa78b59861ad68aa82c99 not found: ID does not exist" Dec 10 20:12:01 crc kubenswrapper[4828]: I1210 20:12:01.214084 4828 scope.go:117] "RemoveContainer" containerID="a3e7739cc24a3123cbff00e8ba834c1a2709be57f525ed1c487b5885bb8c1c9b" Dec 10 20:12:01 crc kubenswrapper[4828]: E1210 20:12:01.214481 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3e7739cc24a3123cbff00e8ba834c1a2709be57f525ed1c487b5885bb8c1c9b\": container with ID starting with a3e7739cc24a3123cbff00e8ba834c1a2709be57f525ed1c487b5885bb8c1c9b not found: ID does not exist" containerID="a3e7739cc24a3123cbff00e8ba834c1a2709be57f525ed1c487b5885bb8c1c9b" Dec 10 20:12:01 crc kubenswrapper[4828]: I1210 20:12:01.214544 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3e7739cc24a3123cbff00e8ba834c1a2709be57f525ed1c487b5885bb8c1c9b"} err="failed to get container status \"a3e7739cc24a3123cbff00e8ba834c1a2709be57f525ed1c487b5885bb8c1c9b\": rpc error: code = NotFound desc = could not find container \"a3e7739cc24a3123cbff00e8ba834c1a2709be57f525ed1c487b5885bb8c1c9b\": container with ID starting with a3e7739cc24a3123cbff00e8ba834c1a2709be57f525ed1c487b5885bb8c1c9b not found: ID does not exist" Dec 10 20:12:01 crc kubenswrapper[4828]: I1210 20:12:01.214588 4828 scope.go:117] "RemoveContainer" containerID="b1f5f22118ecd536fe751b59ebeb5d1d1496a2d095d66cfa5d5a89c865e9ca72" Dec 10 20:12:01 crc kubenswrapper[4828]: E1210 20:12:01.215072 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1f5f22118ecd536fe751b59ebeb5d1d1496a2d095d66cfa5d5a89c865e9ca72\": container with ID starting with b1f5f22118ecd536fe751b59ebeb5d1d1496a2d095d66cfa5d5a89c865e9ca72 not found: ID does not exist" containerID="b1f5f22118ecd536fe751b59ebeb5d1d1496a2d095d66cfa5d5a89c865e9ca72" Dec 10 20:12:01 crc kubenswrapper[4828]: I1210 20:12:01.215104 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1f5f22118ecd536fe751b59ebeb5d1d1496a2d095d66cfa5d5a89c865e9ca72"} err="failed to get container status \"b1f5f22118ecd536fe751b59ebeb5d1d1496a2d095d66cfa5d5a89c865e9ca72\": rpc error: code = NotFound desc = could not find container \"b1f5f22118ecd536fe751b59ebeb5d1d1496a2d095d66cfa5d5a89c865e9ca72\": container with ID starting with b1f5f22118ecd536fe751b59ebeb5d1d1496a2d095d66cfa5d5a89c865e9ca72 not found: ID does not exist" Dec 10 20:12:01 crc kubenswrapper[4828]: I1210 20:12:01.802591 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c93e426a-10a1-46cd-a526-134d4f3d603d" path="/var/lib/kubelet/pods/c93e426a-10a1-46cd-a526-134d4f3d603d/volumes" Dec 10 20:13:39 crc kubenswrapper[4828]: I1210 20:13:39.441867 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5pjnd"] Dec 10 20:13:39 crc kubenswrapper[4828]: E1210 20:13:39.443661 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c93e426a-10a1-46cd-a526-134d4f3d603d" containerName="extract-utilities" Dec 10 20:13:39 crc kubenswrapper[4828]: I1210 20:13:39.443681 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c93e426a-10a1-46cd-a526-134d4f3d603d" containerName="extract-utilities" Dec 10 20:13:39 crc kubenswrapper[4828]: E1210 20:13:39.444009 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c93e426a-10a1-46cd-a526-134d4f3d603d" containerName="extract-content" Dec 10 20:13:39 crc kubenswrapper[4828]: I1210 20:13:39.444022 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c93e426a-10a1-46cd-a526-134d4f3d603d" containerName="extract-content" Dec 10 20:13:39 crc kubenswrapper[4828]: E1210 20:13:39.444056 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c93e426a-10a1-46cd-a526-134d4f3d603d" containerName="registry-server" Dec 10 20:13:39 crc kubenswrapper[4828]: I1210 20:13:39.444065 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c93e426a-10a1-46cd-a526-134d4f3d603d" containerName="registry-server" Dec 10 20:13:39 crc kubenswrapper[4828]: I1210 20:13:39.444502 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c93e426a-10a1-46cd-a526-134d4f3d603d" containerName="registry-server" Dec 10 20:13:39 crc kubenswrapper[4828]: I1210 20:13:39.450747 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5pjnd" Dec 10 20:13:39 crc kubenswrapper[4828]: I1210 20:13:39.465115 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5pjnd"] Dec 10 20:13:39 crc kubenswrapper[4828]: I1210 20:13:39.562524 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f6369db-45b3-45d0-83a5-7425f2f535b6-catalog-content\") pod \"redhat-operators-5pjnd\" (UID: \"3f6369db-45b3-45d0-83a5-7425f2f535b6\") " pod="openshift-marketplace/redhat-operators-5pjnd" Dec 10 20:13:39 crc kubenswrapper[4828]: I1210 20:13:39.562704 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f6369db-45b3-45d0-83a5-7425f2f535b6-utilities\") pod \"redhat-operators-5pjnd\" (UID: \"3f6369db-45b3-45d0-83a5-7425f2f535b6\") " pod="openshift-marketplace/redhat-operators-5pjnd" Dec 10 20:13:39 crc kubenswrapper[4828]: I1210 20:13:39.562761 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sft5\" (UniqueName: \"kubernetes.io/projected/3f6369db-45b3-45d0-83a5-7425f2f535b6-kube-api-access-5sft5\") pod \"redhat-operators-5pjnd\" (UID: \"3f6369db-45b3-45d0-83a5-7425f2f535b6\") " pod="openshift-marketplace/redhat-operators-5pjnd" Dec 10 20:13:39 crc kubenswrapper[4828]: I1210 20:13:39.665874 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f6369db-45b3-45d0-83a5-7425f2f535b6-catalog-content\") pod \"redhat-operators-5pjnd\" (UID: \"3f6369db-45b3-45d0-83a5-7425f2f535b6\") " pod="openshift-marketplace/redhat-operators-5pjnd" Dec 10 20:13:39 crc kubenswrapper[4828]: I1210 20:13:39.665961 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f6369db-45b3-45d0-83a5-7425f2f535b6-utilities\") pod \"redhat-operators-5pjnd\" (UID: \"3f6369db-45b3-45d0-83a5-7425f2f535b6\") " pod="openshift-marketplace/redhat-operators-5pjnd" Dec 10 20:13:39 crc kubenswrapper[4828]: I1210 20:13:39.665994 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sft5\" (UniqueName: \"kubernetes.io/projected/3f6369db-45b3-45d0-83a5-7425f2f535b6-kube-api-access-5sft5\") pod \"redhat-operators-5pjnd\" (UID: \"3f6369db-45b3-45d0-83a5-7425f2f535b6\") " pod="openshift-marketplace/redhat-operators-5pjnd" Dec 10 20:13:39 crc kubenswrapper[4828]: I1210 20:13:39.667335 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f6369db-45b3-45d0-83a5-7425f2f535b6-utilities\") pod \"redhat-operators-5pjnd\" (UID: \"3f6369db-45b3-45d0-83a5-7425f2f535b6\") " pod="openshift-marketplace/redhat-operators-5pjnd" Dec 10 20:13:39 crc kubenswrapper[4828]: I1210 20:13:39.667561 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f6369db-45b3-45d0-83a5-7425f2f535b6-catalog-content\") pod \"redhat-operators-5pjnd\" (UID: \"3f6369db-45b3-45d0-83a5-7425f2f535b6\") " pod="openshift-marketplace/redhat-operators-5pjnd" Dec 10 20:13:39 crc kubenswrapper[4828]: I1210 20:13:39.691168 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sft5\" (UniqueName: \"kubernetes.io/projected/3f6369db-45b3-45d0-83a5-7425f2f535b6-kube-api-access-5sft5\") pod \"redhat-operators-5pjnd\" (UID: \"3f6369db-45b3-45d0-83a5-7425f2f535b6\") " pod="openshift-marketplace/redhat-operators-5pjnd" Dec 10 20:13:39 crc kubenswrapper[4828]: I1210 20:13:39.781674 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5pjnd" Dec 10 20:13:40 crc kubenswrapper[4828]: I1210 20:13:40.592838 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5pjnd"] Dec 10 20:13:41 crc kubenswrapper[4828]: I1210 20:13:41.224146 4828 generic.go:334] "Generic (PLEG): container finished" podID="3f6369db-45b3-45d0-83a5-7425f2f535b6" containerID="bce6b57156172c42bb1c6ac26dbd0dd1ad0890a766cc9c38a9c3f77b16254160" exitCode=0 Dec 10 20:13:41 crc kubenswrapper[4828]: I1210 20:13:41.224612 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pjnd" event={"ID":"3f6369db-45b3-45d0-83a5-7425f2f535b6","Type":"ContainerDied","Data":"bce6b57156172c42bb1c6ac26dbd0dd1ad0890a766cc9c38a9c3f77b16254160"} Dec 10 20:13:41 crc kubenswrapper[4828]: I1210 20:13:41.225047 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pjnd" event={"ID":"3f6369db-45b3-45d0-83a5-7425f2f535b6","Type":"ContainerStarted","Data":"5b71e59a3888051c42d5378a01851fcb0b0cf9523ac96632106f10c9559e1ce3"} Dec 10 20:13:43 crc kubenswrapper[4828]: I1210 20:13:43.249707 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pjnd" event={"ID":"3f6369db-45b3-45d0-83a5-7425f2f535b6","Type":"ContainerStarted","Data":"9fc13402bdda2f592cd72099b269c99ec65f4b22c2cee6fbd040310fc276f479"} Dec 10 20:13:46 crc kubenswrapper[4828]: I1210 20:13:46.283745 4828 generic.go:334] "Generic (PLEG): container finished" podID="3f6369db-45b3-45d0-83a5-7425f2f535b6" containerID="9fc13402bdda2f592cd72099b269c99ec65f4b22c2cee6fbd040310fc276f479" exitCode=0 Dec 10 20:13:46 crc kubenswrapper[4828]: I1210 20:13:46.283839 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pjnd" event={"ID":"3f6369db-45b3-45d0-83a5-7425f2f535b6","Type":"ContainerDied","Data":"9fc13402bdda2f592cd72099b269c99ec65f4b22c2cee6fbd040310fc276f479"} Dec 10 20:13:48 crc kubenswrapper[4828]: I1210 20:13:48.317203 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pjnd" event={"ID":"3f6369db-45b3-45d0-83a5-7425f2f535b6","Type":"ContainerStarted","Data":"539eab4fa21022f1794d7d47c4a9c99691531f43cac9af3039c1c977a90cf837"} Dec 10 20:13:48 crc kubenswrapper[4828]: I1210 20:13:48.345208 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5pjnd" podStartSLOduration=3.6304341449999997 podStartE2EDuration="9.344940478s" podCreationTimestamp="2025-12-10 20:13:39 +0000 UTC" firstStartedPulling="2025-12-10 20:13:41.227023658 +0000 UTC m=+4701.737634663" lastFinishedPulling="2025-12-10 20:13:46.941529991 +0000 UTC m=+4707.452140996" observedRunningTime="2025-12-10 20:13:48.339543734 +0000 UTC m=+4708.850154739" watchObservedRunningTime="2025-12-10 20:13:48.344940478 +0000 UTC m=+4708.855551483" Dec 10 20:13:49 crc kubenswrapper[4828]: I1210 20:13:49.782039 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5pjnd" Dec 10 20:13:49 crc kubenswrapper[4828]: I1210 20:13:49.782699 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5pjnd" Dec 10 20:13:50 crc kubenswrapper[4828]: I1210 20:13:50.842291 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5pjnd" podUID="3f6369db-45b3-45d0-83a5-7425f2f535b6" containerName="registry-server" probeResult="failure" output=< Dec 10 20:13:50 crc kubenswrapper[4828]: timeout: failed to connect service ":50051" within 1s Dec 10 20:13:50 crc kubenswrapper[4828]: > Dec 10 20:14:00 crc kubenswrapper[4828]: I1210 20:14:00.863580 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5pjnd" podUID="3f6369db-45b3-45d0-83a5-7425f2f535b6" containerName="registry-server" probeResult="failure" output=< Dec 10 20:14:00 crc kubenswrapper[4828]: timeout: failed to connect service ":50051" within 1s Dec 10 20:14:00 crc kubenswrapper[4828]: > Dec 10 20:14:09 crc kubenswrapper[4828]: I1210 20:14:09.842777 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5pjnd" Dec 10 20:14:09 crc kubenswrapper[4828]: I1210 20:14:09.902377 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5pjnd" Dec 10 20:14:10 crc kubenswrapper[4828]: I1210 20:14:10.633952 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5pjnd"] Dec 10 20:14:11 crc kubenswrapper[4828]: I1210 20:14:11.599212 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5pjnd" podUID="3f6369db-45b3-45d0-83a5-7425f2f535b6" containerName="registry-server" containerID="cri-o://539eab4fa21022f1794d7d47c4a9c99691531f43cac9af3039c1c977a90cf837" gracePeriod=2 Dec 10 20:14:12 crc kubenswrapper[4828]: I1210 20:14:12.659104 4828 generic.go:334] "Generic (PLEG): container finished" podID="3f6369db-45b3-45d0-83a5-7425f2f535b6" containerID="539eab4fa21022f1794d7d47c4a9c99691531f43cac9af3039c1c977a90cf837" exitCode=0 Dec 10 20:14:12 crc kubenswrapper[4828]: I1210 20:14:12.659989 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pjnd" event={"ID":"3f6369db-45b3-45d0-83a5-7425f2f535b6","Type":"ContainerDied","Data":"539eab4fa21022f1794d7d47c4a9c99691531f43cac9af3039c1c977a90cf837"} Dec 10 20:14:12 crc kubenswrapper[4828]: I1210 20:14:12.660022 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5pjnd" event={"ID":"3f6369db-45b3-45d0-83a5-7425f2f535b6","Type":"ContainerDied","Data":"5b71e59a3888051c42d5378a01851fcb0b0cf9523ac96632106f10c9559e1ce3"} Dec 10 20:14:12 crc kubenswrapper[4828]: I1210 20:14:12.660828 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b71e59a3888051c42d5378a01851fcb0b0cf9523ac96632106f10c9559e1ce3" Dec 10 20:14:12 crc kubenswrapper[4828]: I1210 20:14:12.713147 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5pjnd" Dec 10 20:14:12 crc kubenswrapper[4828]: I1210 20:14:12.840650 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sft5\" (UniqueName: \"kubernetes.io/projected/3f6369db-45b3-45d0-83a5-7425f2f535b6-kube-api-access-5sft5\") pod \"3f6369db-45b3-45d0-83a5-7425f2f535b6\" (UID: \"3f6369db-45b3-45d0-83a5-7425f2f535b6\") " Dec 10 20:14:12 crc kubenswrapper[4828]: I1210 20:14:12.840926 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f6369db-45b3-45d0-83a5-7425f2f535b6-utilities\") pod \"3f6369db-45b3-45d0-83a5-7425f2f535b6\" (UID: \"3f6369db-45b3-45d0-83a5-7425f2f535b6\") " Dec 10 20:14:12 crc kubenswrapper[4828]: I1210 20:14:12.841040 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f6369db-45b3-45d0-83a5-7425f2f535b6-catalog-content\") pod \"3f6369db-45b3-45d0-83a5-7425f2f535b6\" (UID: \"3f6369db-45b3-45d0-83a5-7425f2f535b6\") " Dec 10 20:14:12 crc kubenswrapper[4828]: I1210 20:14:12.842650 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f6369db-45b3-45d0-83a5-7425f2f535b6-utilities" (OuterVolumeSpecName: "utilities") pod "3f6369db-45b3-45d0-83a5-7425f2f535b6" (UID: "3f6369db-45b3-45d0-83a5-7425f2f535b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:14:12 crc kubenswrapper[4828]: I1210 20:14:12.853770 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f6369db-45b3-45d0-83a5-7425f2f535b6-kube-api-access-5sft5" (OuterVolumeSpecName: "kube-api-access-5sft5") pod "3f6369db-45b3-45d0-83a5-7425f2f535b6" (UID: "3f6369db-45b3-45d0-83a5-7425f2f535b6"). InnerVolumeSpecName "kube-api-access-5sft5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:14:12 crc kubenswrapper[4828]: I1210 20:14:12.945355 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sft5\" (UniqueName: \"kubernetes.io/projected/3f6369db-45b3-45d0-83a5-7425f2f535b6-kube-api-access-5sft5\") on node \"crc\" DevicePath \"\"" Dec 10 20:14:12 crc kubenswrapper[4828]: I1210 20:14:12.945936 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f6369db-45b3-45d0-83a5-7425f2f535b6-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:14:12 crc kubenswrapper[4828]: I1210 20:14:12.968227 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f6369db-45b3-45d0-83a5-7425f2f535b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f6369db-45b3-45d0-83a5-7425f2f535b6" (UID: "3f6369db-45b3-45d0-83a5-7425f2f535b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:14:13 crc kubenswrapper[4828]: I1210 20:14:13.048368 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f6369db-45b3-45d0-83a5-7425f2f535b6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:14:13 crc kubenswrapper[4828]: I1210 20:14:13.681382 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5pjnd" Dec 10 20:14:13 crc kubenswrapper[4828]: I1210 20:14:13.734922 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5pjnd"] Dec 10 20:14:13 crc kubenswrapper[4828]: I1210 20:14:13.746270 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5pjnd"] Dec 10 20:14:13 crc kubenswrapper[4828]: I1210 20:14:13.803410 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f6369db-45b3-45d0-83a5-7425f2f535b6" path="/var/lib/kubelet/pods/3f6369db-45b3-45d0-83a5-7425f2f535b6/volumes" Dec 10 20:14:21 crc kubenswrapper[4828]: I1210 20:14:21.230818 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:14:21 crc kubenswrapper[4828]: I1210 20:14:21.231379 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:14:51 crc kubenswrapper[4828]: I1210 20:14:51.230361 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:14:51 crc kubenswrapper[4828]: I1210 20:14:51.232010 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.246395 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj"] Dec 10 20:15:00 crc kubenswrapper[4828]: E1210 20:15:00.247699 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f6369db-45b3-45d0-83a5-7425f2f535b6" containerName="registry-server" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.247716 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f6369db-45b3-45d0-83a5-7425f2f535b6" containerName="registry-server" Dec 10 20:15:00 crc kubenswrapper[4828]: E1210 20:15:00.247815 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f6369db-45b3-45d0-83a5-7425f2f535b6" containerName="extract-utilities" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.247822 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f6369db-45b3-45d0-83a5-7425f2f535b6" containerName="extract-utilities" Dec 10 20:15:00 crc kubenswrapper[4828]: E1210 20:15:00.247837 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f6369db-45b3-45d0-83a5-7425f2f535b6" containerName="extract-content" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.247846 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f6369db-45b3-45d0-83a5-7425f2f535b6" containerName="extract-content" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.248076 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f6369db-45b3-45d0-83a5-7425f2f535b6" containerName="registry-server" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.248955 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.256458 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.256462 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.261057 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj"] Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.327584 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cgfm\" (UniqueName: \"kubernetes.io/projected/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-kube-api-access-8cgfm\") pod \"collect-profiles-29423295-2dxkj\" (UID: \"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.327830 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-config-volume\") pod \"collect-profiles-29423295-2dxkj\" (UID: \"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.328242 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-secret-volume\") pod \"collect-profiles-29423295-2dxkj\" (UID: \"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.432338 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cgfm\" (UniqueName: \"kubernetes.io/projected/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-kube-api-access-8cgfm\") pod \"collect-profiles-29423295-2dxkj\" (UID: \"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.432631 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-config-volume\") pod \"collect-profiles-29423295-2dxkj\" (UID: \"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.432821 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-secret-volume\") pod \"collect-profiles-29423295-2dxkj\" (UID: \"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.434197 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-config-volume\") pod \"collect-profiles-29423295-2dxkj\" (UID: \"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.439275 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-secret-volume\") pod \"collect-profiles-29423295-2dxkj\" (UID: \"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.456137 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cgfm\" (UniqueName: \"kubernetes.io/projected/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-kube-api-access-8cgfm\") pod \"collect-profiles-29423295-2dxkj\" (UID: \"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj" Dec 10 20:15:00 crc kubenswrapper[4828]: I1210 20:15:00.571259 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj" Dec 10 20:15:01 crc kubenswrapper[4828]: I1210 20:15:01.048948 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj"] Dec 10 20:15:01 crc kubenswrapper[4828]: I1210 20:15:01.192453 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj" event={"ID":"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690","Type":"ContainerStarted","Data":"0151f9c0533d1d0e216dcce77d25d8db9f9967aaa60057edd9a3c8df42e693ef"} Dec 10 20:15:02 crc kubenswrapper[4828]: I1210 20:15:02.206156 4828 generic.go:334] "Generic (PLEG): container finished" podID="4dc6e8c1-5c54-4bf7-8424-f9fb3f366690" containerID="4aa0d13dfd1cc63babe1f32820f3f73f84858655a1438f3e75aaef776c96f1b6" exitCode=0 Dec 10 20:15:02 crc kubenswrapper[4828]: I1210 20:15:02.206269 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj" event={"ID":"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690","Type":"ContainerDied","Data":"4aa0d13dfd1cc63babe1f32820f3f73f84858655a1438f3e75aaef776c96f1b6"} Dec 10 20:15:03 crc kubenswrapper[4828]: I1210 20:15:03.842388 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj" Dec 10 20:15:03 crc kubenswrapper[4828]: I1210 20:15:03.929502 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cgfm\" (UniqueName: \"kubernetes.io/projected/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-kube-api-access-8cgfm\") pod \"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690\" (UID: \"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690\") " Dec 10 20:15:03 crc kubenswrapper[4828]: I1210 20:15:03.929903 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-secret-volume\") pod \"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690\" (UID: \"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690\") " Dec 10 20:15:03 crc kubenswrapper[4828]: I1210 20:15:03.930078 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-config-volume\") pod \"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690\" (UID: \"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690\") " Dec 10 20:15:03 crc kubenswrapper[4828]: I1210 20:15:03.930630 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-config-volume" (OuterVolumeSpecName: "config-volume") pod "4dc6e8c1-5c54-4bf7-8424-f9fb3f366690" (UID: "4dc6e8c1-5c54-4bf7-8424-f9fb3f366690"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 20:15:03 crc kubenswrapper[4828]: I1210 20:15:03.931064 4828 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 20:15:03 crc kubenswrapper[4828]: I1210 20:15:03.935609 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-kube-api-access-8cgfm" (OuterVolumeSpecName: "kube-api-access-8cgfm") pod "4dc6e8c1-5c54-4bf7-8424-f9fb3f366690" (UID: "4dc6e8c1-5c54-4bf7-8424-f9fb3f366690"). InnerVolumeSpecName "kube-api-access-8cgfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:15:03 crc kubenswrapper[4828]: I1210 20:15:03.936034 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4dc6e8c1-5c54-4bf7-8424-f9fb3f366690" (UID: "4dc6e8c1-5c54-4bf7-8424-f9fb3f366690"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 20:15:04 crc kubenswrapper[4828]: I1210 20:15:04.032786 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cgfm\" (UniqueName: \"kubernetes.io/projected/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-kube-api-access-8cgfm\") on node \"crc\" DevicePath \"\"" Dec 10 20:15:04 crc kubenswrapper[4828]: I1210 20:15:04.032837 4828 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc6e8c1-5c54-4bf7-8424-f9fb3f366690-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 20:15:04 crc kubenswrapper[4828]: I1210 20:15:04.247518 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj" event={"ID":"4dc6e8c1-5c54-4bf7-8424-f9fb3f366690","Type":"ContainerDied","Data":"0151f9c0533d1d0e216dcce77d25d8db9f9967aaa60057edd9a3c8df42e693ef"} Dec 10 20:15:04 crc kubenswrapper[4828]: I1210 20:15:04.247555 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0151f9c0533d1d0e216dcce77d25d8db9f9967aaa60057edd9a3c8df42e693ef" Dec 10 20:15:04 crc kubenswrapper[4828]: I1210 20:15:04.247572 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423295-2dxkj" Dec 10 20:15:04 crc kubenswrapper[4828]: I1210 20:15:04.945214 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj"] Dec 10 20:15:04 crc kubenswrapper[4828]: I1210 20:15:04.962031 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423250-l5gsj"] Dec 10 20:15:05 crc kubenswrapper[4828]: I1210 20:15:05.805166 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8499d18-c5eb-4373-ab8d-0a20eab63bcc" path="/var/lib/kubelet/pods/e8499d18-c5eb-4373-ab8d-0a20eab63bcc/volumes" Dec 10 20:15:21 crc kubenswrapper[4828]: I1210 20:15:21.230468 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:15:21 crc kubenswrapper[4828]: I1210 20:15:21.231544 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:15:21 crc kubenswrapper[4828]: I1210 20:15:21.231595 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 20:15:21 crc kubenswrapper[4828]: I1210 20:15:21.233057 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7918ddadd3ea00c849d997afdc8c99d59b1ca27ab5e5de92d512e606ec4f36b8"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 20:15:21 crc kubenswrapper[4828]: I1210 20:15:21.233115 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://7918ddadd3ea00c849d997afdc8c99d59b1ca27ab5e5de92d512e606ec4f36b8" gracePeriod=600 Dec 10 20:15:21 crc kubenswrapper[4828]: I1210 20:15:21.461398 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="7918ddadd3ea00c849d997afdc8c99d59b1ca27ab5e5de92d512e606ec4f36b8" exitCode=0 Dec 10 20:15:21 crc kubenswrapper[4828]: I1210 20:15:21.461424 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"7918ddadd3ea00c849d997afdc8c99d59b1ca27ab5e5de92d512e606ec4f36b8"} Dec 10 20:15:21 crc kubenswrapper[4828]: I1210 20:15:21.462535 4828 scope.go:117] "RemoveContainer" containerID="24120c485f9c4b5aa1d1f539258ac1908765629e4e0e5c5b8d44eac0551e5782" Dec 10 20:15:22 crc kubenswrapper[4828]: I1210 20:15:22.474420 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7"} Dec 10 20:15:43 crc kubenswrapper[4828]: I1210 20:15:43.278619 4828 scope.go:117] "RemoveContainer" containerID="6a3d61247abe6456e973ec8131c3bd05e29a51b81157111c3bd315a5a00297eb" Dec 10 20:15:44 crc kubenswrapper[4828]: I1210 20:15:44.185329 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-5bd8d89b5d-tdlk8" podUID="87eba563-d95b-486f-bde7-61d7a764d996" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.94:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 20:17:10 crc kubenswrapper[4828]: I1210 20:17:10.500821 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-55qvx"] Dec 10 20:17:10 crc kubenswrapper[4828]: E1210 20:17:10.515891 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dc6e8c1-5c54-4bf7-8424-f9fb3f366690" containerName="collect-profiles" Dec 10 20:17:10 crc kubenswrapper[4828]: I1210 20:17:10.515922 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dc6e8c1-5c54-4bf7-8424-f9fb3f366690" containerName="collect-profiles" Dec 10 20:17:10 crc kubenswrapper[4828]: I1210 20:17:10.516170 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dc6e8c1-5c54-4bf7-8424-f9fb3f366690" containerName="collect-profiles" Dec 10 20:17:10 crc kubenswrapper[4828]: I1210 20:17:10.518444 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-55qvx" Dec 10 20:17:10 crc kubenswrapper[4828]: I1210 20:17:10.539065 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-55qvx"] Dec 10 20:17:10 crc kubenswrapper[4828]: I1210 20:17:10.693183 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60469732-5c75-4ff5-b0e4-752b73a0af26-utilities\") pod \"redhat-marketplace-55qvx\" (UID: \"60469732-5c75-4ff5-b0e4-752b73a0af26\") " pod="openshift-marketplace/redhat-marketplace-55qvx" Dec 10 20:17:10 crc kubenswrapper[4828]: I1210 20:17:10.693250 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60469732-5c75-4ff5-b0e4-752b73a0af26-catalog-content\") pod \"redhat-marketplace-55qvx\" (UID: \"60469732-5c75-4ff5-b0e4-752b73a0af26\") " pod="openshift-marketplace/redhat-marketplace-55qvx" Dec 10 20:17:10 crc kubenswrapper[4828]: I1210 20:17:10.693285 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkwjt\" (UniqueName: \"kubernetes.io/projected/60469732-5c75-4ff5-b0e4-752b73a0af26-kube-api-access-fkwjt\") pod \"redhat-marketplace-55qvx\" (UID: \"60469732-5c75-4ff5-b0e4-752b73a0af26\") " pod="openshift-marketplace/redhat-marketplace-55qvx" Dec 10 20:17:10 crc kubenswrapper[4828]: I1210 20:17:10.794458 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60469732-5c75-4ff5-b0e4-752b73a0af26-utilities\") pod \"redhat-marketplace-55qvx\" (UID: \"60469732-5c75-4ff5-b0e4-752b73a0af26\") " pod="openshift-marketplace/redhat-marketplace-55qvx" Dec 10 20:17:10 crc kubenswrapper[4828]: I1210 20:17:10.794515 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60469732-5c75-4ff5-b0e4-752b73a0af26-catalog-content\") pod \"redhat-marketplace-55qvx\" (UID: \"60469732-5c75-4ff5-b0e4-752b73a0af26\") " pod="openshift-marketplace/redhat-marketplace-55qvx" Dec 10 20:17:10 crc kubenswrapper[4828]: I1210 20:17:10.794544 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkwjt\" (UniqueName: \"kubernetes.io/projected/60469732-5c75-4ff5-b0e4-752b73a0af26-kube-api-access-fkwjt\") pod \"redhat-marketplace-55qvx\" (UID: \"60469732-5c75-4ff5-b0e4-752b73a0af26\") " pod="openshift-marketplace/redhat-marketplace-55qvx" Dec 10 20:17:10 crc kubenswrapper[4828]: I1210 20:17:10.794893 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60469732-5c75-4ff5-b0e4-752b73a0af26-utilities\") pod \"redhat-marketplace-55qvx\" (UID: \"60469732-5c75-4ff5-b0e4-752b73a0af26\") " pod="openshift-marketplace/redhat-marketplace-55qvx" Dec 10 20:17:10 crc kubenswrapper[4828]: I1210 20:17:10.795088 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60469732-5c75-4ff5-b0e4-752b73a0af26-catalog-content\") pod \"redhat-marketplace-55qvx\" (UID: \"60469732-5c75-4ff5-b0e4-752b73a0af26\") " pod="openshift-marketplace/redhat-marketplace-55qvx" Dec 10 20:17:10 crc kubenswrapper[4828]: I1210 20:17:10.814410 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkwjt\" (UniqueName: \"kubernetes.io/projected/60469732-5c75-4ff5-b0e4-752b73a0af26-kube-api-access-fkwjt\") pod \"redhat-marketplace-55qvx\" (UID: \"60469732-5c75-4ff5-b0e4-752b73a0af26\") " pod="openshift-marketplace/redhat-marketplace-55qvx" Dec 10 20:17:10 crc kubenswrapper[4828]: I1210 20:17:10.848658 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-55qvx" Dec 10 20:17:11 crc kubenswrapper[4828]: I1210 20:17:11.342018 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-55qvx"] Dec 10 20:17:11 crc kubenswrapper[4828]: W1210 20:17:11.344574 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60469732_5c75_4ff5_b0e4_752b73a0af26.slice/crio-5e20bd40d0dd906ce611db0330eda13f08f3cc189a8065adb0bf8b29b321f49b WatchSource:0}: Error finding container 5e20bd40d0dd906ce611db0330eda13f08f3cc189a8065adb0bf8b29b321f49b: Status 404 returned error can't find the container with id 5e20bd40d0dd906ce611db0330eda13f08f3cc189a8065adb0bf8b29b321f49b Dec 10 20:17:11 crc kubenswrapper[4828]: I1210 20:17:11.636310 4828 generic.go:334] "Generic (PLEG): container finished" podID="60469732-5c75-4ff5-b0e4-752b73a0af26" containerID="c32cd4aee5d3368ae91f809f8cc1a21dc6c5f6c3482f3cd99a886af8f9ee3ee0" exitCode=0 Dec 10 20:17:11 crc kubenswrapper[4828]: I1210 20:17:11.636467 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-55qvx" event={"ID":"60469732-5c75-4ff5-b0e4-752b73a0af26","Type":"ContainerDied","Data":"c32cd4aee5d3368ae91f809f8cc1a21dc6c5f6c3482f3cd99a886af8f9ee3ee0"} Dec 10 20:17:11 crc kubenswrapper[4828]: I1210 20:17:11.636598 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-55qvx" event={"ID":"60469732-5c75-4ff5-b0e4-752b73a0af26","Type":"ContainerStarted","Data":"5e20bd40d0dd906ce611db0330eda13f08f3cc189a8065adb0bf8b29b321f49b"} Dec 10 20:17:11 crc kubenswrapper[4828]: I1210 20:17:11.641548 4828 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 20:17:15 crc kubenswrapper[4828]: I1210 20:17:15.682150 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-55qvx" event={"ID":"60469732-5c75-4ff5-b0e4-752b73a0af26","Type":"ContainerStarted","Data":"79aa6608699f6b5b659145240c33a93962e6cbaabd9ad7f6709f5a7fd8dd7964"} Dec 10 20:17:16 crc kubenswrapper[4828]: I1210 20:17:16.693070 4828 generic.go:334] "Generic (PLEG): container finished" podID="60469732-5c75-4ff5-b0e4-752b73a0af26" containerID="79aa6608699f6b5b659145240c33a93962e6cbaabd9ad7f6709f5a7fd8dd7964" exitCode=0 Dec 10 20:17:16 crc kubenswrapper[4828]: I1210 20:17:16.693189 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-55qvx" event={"ID":"60469732-5c75-4ff5-b0e4-752b73a0af26","Type":"ContainerDied","Data":"79aa6608699f6b5b659145240c33a93962e6cbaabd9ad7f6709f5a7fd8dd7964"} Dec 10 20:17:17 crc kubenswrapper[4828]: I1210 20:17:17.705912 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-55qvx" event={"ID":"60469732-5c75-4ff5-b0e4-752b73a0af26","Type":"ContainerStarted","Data":"687268f1908cae12b0143593ce4c30d1f2fa3252357513bb0cd1188c00af1c92"} Dec 10 20:17:17 crc kubenswrapper[4828]: I1210 20:17:17.730507 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-55qvx" podStartSLOduration=2.06021709 podStartE2EDuration="7.730205292s" podCreationTimestamp="2025-12-10 20:17:10 +0000 UTC" firstStartedPulling="2025-12-10 20:17:11.639631166 +0000 UTC m=+4912.150242171" lastFinishedPulling="2025-12-10 20:17:17.309619358 +0000 UTC m=+4917.820230373" observedRunningTime="2025-12-10 20:17:17.727644934 +0000 UTC m=+4918.238255979" watchObservedRunningTime="2025-12-10 20:17:17.730205292 +0000 UTC m=+4918.240816297" Dec 10 20:17:20 crc kubenswrapper[4828]: I1210 20:17:20.849631 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-55qvx" Dec 10 20:17:20 crc kubenswrapper[4828]: I1210 20:17:20.850397 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-55qvx" Dec 10 20:17:20 crc kubenswrapper[4828]: I1210 20:17:20.914665 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-55qvx" Dec 10 20:17:21 crc kubenswrapper[4828]: I1210 20:17:21.230400 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:17:21 crc kubenswrapper[4828]: I1210 20:17:21.230789 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:17:27 crc kubenswrapper[4828]: I1210 20:17:27.772278 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-44fpz"] Dec 10 20:17:27 crc kubenswrapper[4828]: I1210 20:17:27.775835 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-44fpz" Dec 10 20:17:27 crc kubenswrapper[4828]: I1210 20:17:27.806301 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-44fpz"] Dec 10 20:17:27 crc kubenswrapper[4828]: I1210 20:17:27.899865 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-utilities\") pod \"community-operators-44fpz\" (UID: \"f59f9dc7-df56-4ff1-bf33-0b8c172854f3\") " pod="openshift-marketplace/community-operators-44fpz" Dec 10 20:17:27 crc kubenswrapper[4828]: I1210 20:17:27.900007 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-catalog-content\") pod \"community-operators-44fpz\" (UID: \"f59f9dc7-df56-4ff1-bf33-0b8c172854f3\") " pod="openshift-marketplace/community-operators-44fpz" Dec 10 20:17:27 crc kubenswrapper[4828]: I1210 20:17:27.900041 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7bvf\" (UniqueName: \"kubernetes.io/projected/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-kube-api-access-m7bvf\") pod \"community-operators-44fpz\" (UID: \"f59f9dc7-df56-4ff1-bf33-0b8c172854f3\") " pod="openshift-marketplace/community-operators-44fpz" Dec 10 20:17:28 crc kubenswrapper[4828]: I1210 20:17:28.002297 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-catalog-content\") pod \"community-operators-44fpz\" (UID: \"f59f9dc7-df56-4ff1-bf33-0b8c172854f3\") " pod="openshift-marketplace/community-operators-44fpz" Dec 10 20:17:28 crc kubenswrapper[4828]: I1210 20:17:28.002362 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7bvf\" (UniqueName: \"kubernetes.io/projected/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-kube-api-access-m7bvf\") pod \"community-operators-44fpz\" (UID: \"f59f9dc7-df56-4ff1-bf33-0b8c172854f3\") " pod="openshift-marketplace/community-operators-44fpz" Dec 10 20:17:28 crc kubenswrapper[4828]: I1210 20:17:28.002537 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-utilities\") pod \"community-operators-44fpz\" (UID: \"f59f9dc7-df56-4ff1-bf33-0b8c172854f3\") " pod="openshift-marketplace/community-operators-44fpz" Dec 10 20:17:28 crc kubenswrapper[4828]: I1210 20:17:28.002815 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-catalog-content\") pod \"community-operators-44fpz\" (UID: \"f59f9dc7-df56-4ff1-bf33-0b8c172854f3\") " pod="openshift-marketplace/community-operators-44fpz" Dec 10 20:17:28 crc kubenswrapper[4828]: I1210 20:17:28.003057 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-utilities\") pod \"community-operators-44fpz\" (UID: \"f59f9dc7-df56-4ff1-bf33-0b8c172854f3\") " pod="openshift-marketplace/community-operators-44fpz" Dec 10 20:17:28 crc kubenswrapper[4828]: I1210 20:17:28.020508 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7bvf\" (UniqueName: \"kubernetes.io/projected/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-kube-api-access-m7bvf\") pod \"community-operators-44fpz\" (UID: \"f59f9dc7-df56-4ff1-bf33-0b8c172854f3\") " pod="openshift-marketplace/community-operators-44fpz" Dec 10 20:17:28 crc kubenswrapper[4828]: I1210 20:17:28.099912 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-44fpz" Dec 10 20:17:28 crc kubenswrapper[4828]: I1210 20:17:28.620681 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-44fpz"] Dec 10 20:17:28 crc kubenswrapper[4828]: W1210 20:17:28.626599 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf59f9dc7_df56_4ff1_bf33_0b8c172854f3.slice/crio-99e1239256f313c0d4c0e6e2d2e0a99bd129468da03822af3701fa76272e0bb2 WatchSource:0}: Error finding container 99e1239256f313c0d4c0e6e2d2e0a99bd129468da03822af3701fa76272e0bb2: Status 404 returned error can't find the container with id 99e1239256f313c0d4c0e6e2d2e0a99bd129468da03822af3701fa76272e0bb2 Dec 10 20:17:28 crc kubenswrapper[4828]: I1210 20:17:28.859233 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-44fpz" event={"ID":"f59f9dc7-df56-4ff1-bf33-0b8c172854f3","Type":"ContainerStarted","Data":"acd95b8a6792f68deef6e83c0093c694214ae672f20077c3e0c4e08ff21d0c41"} Dec 10 20:17:28 crc kubenswrapper[4828]: I1210 20:17:28.859555 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-44fpz" event={"ID":"f59f9dc7-df56-4ff1-bf33-0b8c172854f3","Type":"ContainerStarted","Data":"99e1239256f313c0d4c0e6e2d2e0a99bd129468da03822af3701fa76272e0bb2"} Dec 10 20:17:29 crc kubenswrapper[4828]: I1210 20:17:29.871491 4828 generic.go:334] "Generic (PLEG): container finished" podID="f59f9dc7-df56-4ff1-bf33-0b8c172854f3" containerID="acd95b8a6792f68deef6e83c0093c694214ae672f20077c3e0c4e08ff21d0c41" exitCode=0 Dec 10 20:17:29 crc kubenswrapper[4828]: I1210 20:17:29.871596 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-44fpz" event={"ID":"f59f9dc7-df56-4ff1-bf33-0b8c172854f3","Type":"ContainerDied","Data":"acd95b8a6792f68deef6e83c0093c694214ae672f20077c3e0c4e08ff21d0c41"} Dec 10 20:17:30 crc kubenswrapper[4828]: I1210 20:17:30.897431 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-44fpz" event={"ID":"f59f9dc7-df56-4ff1-bf33-0b8c172854f3","Type":"ContainerStarted","Data":"6a70a7a54986a353927960ac3377c1d698453ec19763353bf4d0910d99e65031"} Dec 10 20:17:30 crc kubenswrapper[4828]: I1210 20:17:30.913990 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-55qvx" Dec 10 20:17:31 crc kubenswrapper[4828]: I1210 20:17:31.910150 4828 generic.go:334] "Generic (PLEG): container finished" podID="f59f9dc7-df56-4ff1-bf33-0b8c172854f3" containerID="6a70a7a54986a353927960ac3377c1d698453ec19763353bf4d0910d99e65031" exitCode=0 Dec 10 20:17:31 crc kubenswrapper[4828]: I1210 20:17:31.910218 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-44fpz" event={"ID":"f59f9dc7-df56-4ff1-bf33-0b8c172854f3","Type":"ContainerDied","Data":"6a70a7a54986a353927960ac3377c1d698453ec19763353bf4d0910d99e65031"} Dec 10 20:17:32 crc kubenswrapper[4828]: I1210 20:17:32.934202 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-44fpz" event={"ID":"f59f9dc7-df56-4ff1-bf33-0b8c172854f3","Type":"ContainerStarted","Data":"7f9902cd3754ab6123949dbbc8c6b4b599d3b86b376dfd4422f86b28beb13e38"} Dec 10 20:17:32 crc kubenswrapper[4828]: I1210 20:17:32.954037 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-44fpz" podStartSLOduration=3.378651458 podStartE2EDuration="5.9540177s" podCreationTimestamp="2025-12-10 20:17:27 +0000 UTC" firstStartedPulling="2025-12-10 20:17:29.873537798 +0000 UTC m=+4930.384148803" lastFinishedPulling="2025-12-10 20:17:32.44890403 +0000 UTC m=+4932.959515045" observedRunningTime="2025-12-10 20:17:32.951235256 +0000 UTC m=+4933.461846271" watchObservedRunningTime="2025-12-10 20:17:32.9540177 +0000 UTC m=+4933.464628715" Dec 10 20:17:33 crc kubenswrapper[4828]: I1210 20:17:33.350447 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-55qvx"] Dec 10 20:17:33 crc kubenswrapper[4828]: I1210 20:17:33.350683 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-55qvx" podUID="60469732-5c75-4ff5-b0e4-752b73a0af26" containerName="registry-server" containerID="cri-o://687268f1908cae12b0143593ce4c30d1f2fa3252357513bb0cd1188c00af1c92" gracePeriod=2 Dec 10 20:17:33 crc kubenswrapper[4828]: I1210 20:17:33.929109 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-55qvx" Dec 10 20:17:33 crc kubenswrapper[4828]: I1210 20:17:33.945480 4828 generic.go:334] "Generic (PLEG): container finished" podID="60469732-5c75-4ff5-b0e4-752b73a0af26" containerID="687268f1908cae12b0143593ce4c30d1f2fa3252357513bb0cd1188c00af1c92" exitCode=0 Dec 10 20:17:33 crc kubenswrapper[4828]: I1210 20:17:33.945581 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-55qvx" event={"ID":"60469732-5c75-4ff5-b0e4-752b73a0af26","Type":"ContainerDied","Data":"687268f1908cae12b0143593ce4c30d1f2fa3252357513bb0cd1188c00af1c92"} Dec 10 20:17:33 crc kubenswrapper[4828]: I1210 20:17:33.945671 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-55qvx" event={"ID":"60469732-5c75-4ff5-b0e4-752b73a0af26","Type":"ContainerDied","Data":"5e20bd40d0dd906ce611db0330eda13f08f3cc189a8065adb0bf8b29b321f49b"} Dec 10 20:17:33 crc kubenswrapper[4828]: I1210 20:17:33.945692 4828 scope.go:117] "RemoveContainer" containerID="687268f1908cae12b0143593ce4c30d1f2fa3252357513bb0cd1188c00af1c92" Dec 10 20:17:33 crc kubenswrapper[4828]: I1210 20:17:33.946425 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-55qvx" Dec 10 20:17:33 crc kubenswrapper[4828]: I1210 20:17:33.978533 4828 scope.go:117] "RemoveContainer" containerID="79aa6608699f6b5b659145240c33a93962e6cbaabd9ad7f6709f5a7fd8dd7964" Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.011270 4828 scope.go:117] "RemoveContainer" containerID="c32cd4aee5d3368ae91f809f8cc1a21dc6c5f6c3482f3cd99a886af8f9ee3ee0" Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.052994 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkwjt\" (UniqueName: \"kubernetes.io/projected/60469732-5c75-4ff5-b0e4-752b73a0af26-kube-api-access-fkwjt\") pod \"60469732-5c75-4ff5-b0e4-752b73a0af26\" (UID: \"60469732-5c75-4ff5-b0e4-752b73a0af26\") " Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.053153 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60469732-5c75-4ff5-b0e4-752b73a0af26-utilities\") pod \"60469732-5c75-4ff5-b0e4-752b73a0af26\" (UID: \"60469732-5c75-4ff5-b0e4-752b73a0af26\") " Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.053237 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60469732-5c75-4ff5-b0e4-752b73a0af26-catalog-content\") pod \"60469732-5c75-4ff5-b0e4-752b73a0af26\" (UID: \"60469732-5c75-4ff5-b0e4-752b73a0af26\") " Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.054364 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60469732-5c75-4ff5-b0e4-752b73a0af26-utilities" (OuterVolumeSpecName: "utilities") pod "60469732-5c75-4ff5-b0e4-752b73a0af26" (UID: "60469732-5c75-4ff5-b0e4-752b73a0af26"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.060870 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60469732-5c75-4ff5-b0e4-752b73a0af26-kube-api-access-fkwjt" (OuterVolumeSpecName: "kube-api-access-fkwjt") pod "60469732-5c75-4ff5-b0e4-752b73a0af26" (UID: "60469732-5c75-4ff5-b0e4-752b73a0af26"). InnerVolumeSpecName "kube-api-access-fkwjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.071032 4828 scope.go:117] "RemoveContainer" containerID="687268f1908cae12b0143593ce4c30d1f2fa3252357513bb0cd1188c00af1c92" Dec 10 20:17:34 crc kubenswrapper[4828]: E1210 20:17:34.072046 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"687268f1908cae12b0143593ce4c30d1f2fa3252357513bb0cd1188c00af1c92\": container with ID starting with 687268f1908cae12b0143593ce4c30d1f2fa3252357513bb0cd1188c00af1c92 not found: ID does not exist" containerID="687268f1908cae12b0143593ce4c30d1f2fa3252357513bb0cd1188c00af1c92" Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.072107 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"687268f1908cae12b0143593ce4c30d1f2fa3252357513bb0cd1188c00af1c92"} err="failed to get container status \"687268f1908cae12b0143593ce4c30d1f2fa3252357513bb0cd1188c00af1c92\": rpc error: code = NotFound desc = could not find container \"687268f1908cae12b0143593ce4c30d1f2fa3252357513bb0cd1188c00af1c92\": container with ID starting with 687268f1908cae12b0143593ce4c30d1f2fa3252357513bb0cd1188c00af1c92 not found: ID does not exist" Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.072133 4828 scope.go:117] "RemoveContainer" containerID="79aa6608699f6b5b659145240c33a93962e6cbaabd9ad7f6709f5a7fd8dd7964" Dec 10 20:17:34 crc kubenswrapper[4828]: E1210 20:17:34.073110 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79aa6608699f6b5b659145240c33a93962e6cbaabd9ad7f6709f5a7fd8dd7964\": container with ID starting with 79aa6608699f6b5b659145240c33a93962e6cbaabd9ad7f6709f5a7fd8dd7964 not found: ID does not exist" containerID="79aa6608699f6b5b659145240c33a93962e6cbaabd9ad7f6709f5a7fd8dd7964" Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.073136 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79aa6608699f6b5b659145240c33a93962e6cbaabd9ad7f6709f5a7fd8dd7964"} err="failed to get container status \"79aa6608699f6b5b659145240c33a93962e6cbaabd9ad7f6709f5a7fd8dd7964\": rpc error: code = NotFound desc = could not find container \"79aa6608699f6b5b659145240c33a93962e6cbaabd9ad7f6709f5a7fd8dd7964\": container with ID starting with 79aa6608699f6b5b659145240c33a93962e6cbaabd9ad7f6709f5a7fd8dd7964 not found: ID does not exist" Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.073154 4828 scope.go:117] "RemoveContainer" containerID="c32cd4aee5d3368ae91f809f8cc1a21dc6c5f6c3482f3cd99a886af8f9ee3ee0" Dec 10 20:17:34 crc kubenswrapper[4828]: E1210 20:17:34.073447 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c32cd4aee5d3368ae91f809f8cc1a21dc6c5f6c3482f3cd99a886af8f9ee3ee0\": container with ID starting with c32cd4aee5d3368ae91f809f8cc1a21dc6c5f6c3482f3cd99a886af8f9ee3ee0 not found: ID does not exist" containerID="c32cd4aee5d3368ae91f809f8cc1a21dc6c5f6c3482f3cd99a886af8f9ee3ee0" Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.073476 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c32cd4aee5d3368ae91f809f8cc1a21dc6c5f6c3482f3cd99a886af8f9ee3ee0"} err="failed to get container status \"c32cd4aee5d3368ae91f809f8cc1a21dc6c5f6c3482f3cd99a886af8f9ee3ee0\": rpc error: code = NotFound desc = could not find container \"c32cd4aee5d3368ae91f809f8cc1a21dc6c5f6c3482f3cd99a886af8f9ee3ee0\": container with ID starting with c32cd4aee5d3368ae91f809f8cc1a21dc6c5f6c3482f3cd99a886af8f9ee3ee0 not found: ID does not exist" Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.074930 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60469732-5c75-4ff5-b0e4-752b73a0af26-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "60469732-5c75-4ff5-b0e4-752b73a0af26" (UID: "60469732-5c75-4ff5-b0e4-752b73a0af26"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.156808 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkwjt\" (UniqueName: \"kubernetes.io/projected/60469732-5c75-4ff5-b0e4-752b73a0af26-kube-api-access-fkwjt\") on node \"crc\" DevicePath \"\"" Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.156840 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/60469732-5c75-4ff5-b0e4-752b73a0af26-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.156850 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/60469732-5c75-4ff5-b0e4-752b73a0af26-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.287848 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-55qvx"] Dec 10 20:17:34 crc kubenswrapper[4828]: I1210 20:17:34.320158 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-55qvx"] Dec 10 20:17:35 crc kubenswrapper[4828]: I1210 20:17:35.806008 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60469732-5c75-4ff5-b0e4-752b73a0af26" path="/var/lib/kubelet/pods/60469732-5c75-4ff5-b0e4-752b73a0af26/volumes" Dec 10 20:17:38 crc kubenswrapper[4828]: I1210 20:17:38.100958 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-44fpz" Dec 10 20:17:38 crc kubenswrapper[4828]: I1210 20:17:38.101512 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-44fpz" Dec 10 20:17:38 crc kubenswrapper[4828]: I1210 20:17:38.153516 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-44fpz" Dec 10 20:17:39 crc kubenswrapper[4828]: I1210 20:17:39.045467 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-44fpz" Dec 10 20:17:39 crc kubenswrapper[4828]: I1210 20:17:39.354271 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-44fpz"] Dec 10 20:17:41 crc kubenswrapper[4828]: I1210 20:17:41.027925 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-44fpz" podUID="f59f9dc7-df56-4ff1-bf33-0b8c172854f3" containerName="registry-server" containerID="cri-o://7f9902cd3754ab6123949dbbc8c6b4b599d3b86b376dfd4422f86b28beb13e38" gracePeriod=2 Dec 10 20:17:41 crc kubenswrapper[4828]: I1210 20:17:41.712101 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-44fpz" Dec 10 20:17:41 crc kubenswrapper[4828]: I1210 20:17:41.846690 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7bvf\" (UniqueName: \"kubernetes.io/projected/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-kube-api-access-m7bvf\") pod \"f59f9dc7-df56-4ff1-bf33-0b8c172854f3\" (UID: \"f59f9dc7-df56-4ff1-bf33-0b8c172854f3\") " Dec 10 20:17:41 crc kubenswrapper[4828]: I1210 20:17:41.846835 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-catalog-content\") pod \"f59f9dc7-df56-4ff1-bf33-0b8c172854f3\" (UID: \"f59f9dc7-df56-4ff1-bf33-0b8c172854f3\") " Dec 10 20:17:41 crc kubenswrapper[4828]: I1210 20:17:41.846948 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-utilities\") pod \"f59f9dc7-df56-4ff1-bf33-0b8c172854f3\" (UID: \"f59f9dc7-df56-4ff1-bf33-0b8c172854f3\") " Dec 10 20:17:41 crc kubenswrapper[4828]: I1210 20:17:41.847556 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-utilities" (OuterVolumeSpecName: "utilities") pod "f59f9dc7-df56-4ff1-bf33-0b8c172854f3" (UID: "f59f9dc7-df56-4ff1-bf33-0b8c172854f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:17:41 crc kubenswrapper[4828]: I1210 20:17:41.852487 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-kube-api-access-m7bvf" (OuterVolumeSpecName: "kube-api-access-m7bvf") pod "f59f9dc7-df56-4ff1-bf33-0b8c172854f3" (UID: "f59f9dc7-df56-4ff1-bf33-0b8c172854f3"). InnerVolumeSpecName "kube-api-access-m7bvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:17:41 crc kubenswrapper[4828]: I1210 20:17:41.892119 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f59f9dc7-df56-4ff1-bf33-0b8c172854f3" (UID: "f59f9dc7-df56-4ff1-bf33-0b8c172854f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:17:41 crc kubenswrapper[4828]: I1210 20:17:41.950497 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7bvf\" (UniqueName: \"kubernetes.io/projected/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-kube-api-access-m7bvf\") on node \"crc\" DevicePath \"\"" Dec 10 20:17:41 crc kubenswrapper[4828]: I1210 20:17:41.950677 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:17:41 crc kubenswrapper[4828]: I1210 20:17:41.950703 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f59f9dc7-df56-4ff1-bf33-0b8c172854f3-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:17:42 crc kubenswrapper[4828]: I1210 20:17:42.041890 4828 generic.go:334] "Generic (PLEG): container finished" podID="f59f9dc7-df56-4ff1-bf33-0b8c172854f3" containerID="7f9902cd3754ab6123949dbbc8c6b4b599d3b86b376dfd4422f86b28beb13e38" exitCode=0 Dec 10 20:17:42 crc kubenswrapper[4828]: I1210 20:17:42.041955 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-44fpz" event={"ID":"f59f9dc7-df56-4ff1-bf33-0b8c172854f3","Type":"ContainerDied","Data":"7f9902cd3754ab6123949dbbc8c6b4b599d3b86b376dfd4422f86b28beb13e38"} Dec 10 20:17:42 crc kubenswrapper[4828]: I1210 20:17:42.041987 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-44fpz" Dec 10 20:17:42 crc kubenswrapper[4828]: I1210 20:17:42.042013 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-44fpz" event={"ID":"f59f9dc7-df56-4ff1-bf33-0b8c172854f3","Type":"ContainerDied","Data":"99e1239256f313c0d4c0e6e2d2e0a99bd129468da03822af3701fa76272e0bb2"} Dec 10 20:17:42 crc kubenswrapper[4828]: I1210 20:17:42.042035 4828 scope.go:117] "RemoveContainer" containerID="7f9902cd3754ab6123949dbbc8c6b4b599d3b86b376dfd4422f86b28beb13e38" Dec 10 20:17:42 crc kubenswrapper[4828]: I1210 20:17:42.075308 4828 scope.go:117] "RemoveContainer" containerID="6a70a7a54986a353927960ac3377c1d698453ec19763353bf4d0910d99e65031" Dec 10 20:17:42 crc kubenswrapper[4828]: I1210 20:17:42.110811 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-44fpz"] Dec 10 20:17:42 crc kubenswrapper[4828]: I1210 20:17:42.120278 4828 scope.go:117] "RemoveContainer" containerID="acd95b8a6792f68deef6e83c0093c694214ae672f20077c3e0c4e08ff21d0c41" Dec 10 20:17:42 crc kubenswrapper[4828]: I1210 20:17:42.124174 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-44fpz"] Dec 10 20:17:42 crc kubenswrapper[4828]: I1210 20:17:42.156757 4828 scope.go:117] "RemoveContainer" containerID="7f9902cd3754ab6123949dbbc8c6b4b599d3b86b376dfd4422f86b28beb13e38" Dec 10 20:17:42 crc kubenswrapper[4828]: E1210 20:17:42.158619 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f9902cd3754ab6123949dbbc8c6b4b599d3b86b376dfd4422f86b28beb13e38\": container with ID starting with 7f9902cd3754ab6123949dbbc8c6b4b599d3b86b376dfd4422f86b28beb13e38 not found: ID does not exist" containerID="7f9902cd3754ab6123949dbbc8c6b4b599d3b86b376dfd4422f86b28beb13e38" Dec 10 20:17:42 crc kubenswrapper[4828]: I1210 20:17:42.158666 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f9902cd3754ab6123949dbbc8c6b4b599d3b86b376dfd4422f86b28beb13e38"} err="failed to get container status \"7f9902cd3754ab6123949dbbc8c6b4b599d3b86b376dfd4422f86b28beb13e38\": rpc error: code = NotFound desc = could not find container \"7f9902cd3754ab6123949dbbc8c6b4b599d3b86b376dfd4422f86b28beb13e38\": container with ID starting with 7f9902cd3754ab6123949dbbc8c6b4b599d3b86b376dfd4422f86b28beb13e38 not found: ID does not exist" Dec 10 20:17:42 crc kubenswrapper[4828]: I1210 20:17:42.158696 4828 scope.go:117] "RemoveContainer" containerID="6a70a7a54986a353927960ac3377c1d698453ec19763353bf4d0910d99e65031" Dec 10 20:17:42 crc kubenswrapper[4828]: E1210 20:17:42.163695 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a70a7a54986a353927960ac3377c1d698453ec19763353bf4d0910d99e65031\": container with ID starting with 6a70a7a54986a353927960ac3377c1d698453ec19763353bf4d0910d99e65031 not found: ID does not exist" containerID="6a70a7a54986a353927960ac3377c1d698453ec19763353bf4d0910d99e65031" Dec 10 20:17:42 crc kubenswrapper[4828]: I1210 20:17:42.163750 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a70a7a54986a353927960ac3377c1d698453ec19763353bf4d0910d99e65031"} err="failed to get container status \"6a70a7a54986a353927960ac3377c1d698453ec19763353bf4d0910d99e65031\": rpc error: code = NotFound desc = could not find container \"6a70a7a54986a353927960ac3377c1d698453ec19763353bf4d0910d99e65031\": container with ID starting with 6a70a7a54986a353927960ac3377c1d698453ec19763353bf4d0910d99e65031 not found: ID does not exist" Dec 10 20:17:42 crc kubenswrapper[4828]: I1210 20:17:42.163804 4828 scope.go:117] "RemoveContainer" containerID="acd95b8a6792f68deef6e83c0093c694214ae672f20077c3e0c4e08ff21d0c41" Dec 10 20:17:42 crc kubenswrapper[4828]: E1210 20:17:42.164237 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acd95b8a6792f68deef6e83c0093c694214ae672f20077c3e0c4e08ff21d0c41\": container with ID starting with acd95b8a6792f68deef6e83c0093c694214ae672f20077c3e0c4e08ff21d0c41 not found: ID does not exist" containerID="acd95b8a6792f68deef6e83c0093c694214ae672f20077c3e0c4e08ff21d0c41" Dec 10 20:17:42 crc kubenswrapper[4828]: I1210 20:17:42.164279 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acd95b8a6792f68deef6e83c0093c694214ae672f20077c3e0c4e08ff21d0c41"} err="failed to get container status \"acd95b8a6792f68deef6e83c0093c694214ae672f20077c3e0c4e08ff21d0c41\": rpc error: code = NotFound desc = could not find container \"acd95b8a6792f68deef6e83c0093c694214ae672f20077c3e0c4e08ff21d0c41\": container with ID starting with acd95b8a6792f68deef6e83c0093c694214ae672f20077c3e0c4e08ff21d0c41 not found: ID does not exist" Dec 10 20:17:43 crc kubenswrapper[4828]: I1210 20:17:43.801995 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f59f9dc7-df56-4ff1-bf33-0b8c172854f3" path="/var/lib/kubelet/pods/f59f9dc7-df56-4ff1-bf33-0b8c172854f3/volumes" Dec 10 20:17:51 crc kubenswrapper[4828]: I1210 20:17:51.230306 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:17:51 crc kubenswrapper[4828]: I1210 20:17:51.230819 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:18:21 crc kubenswrapper[4828]: I1210 20:18:21.230193 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:18:21 crc kubenswrapper[4828]: I1210 20:18:21.230947 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:18:21 crc kubenswrapper[4828]: I1210 20:18:21.231029 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 20:18:21 crc kubenswrapper[4828]: I1210 20:18:21.232172 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 20:18:21 crc kubenswrapper[4828]: I1210 20:18:21.232253 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" gracePeriod=600 Dec 10 20:18:21 crc kubenswrapper[4828]: E1210 20:18:21.354014 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:18:21 crc kubenswrapper[4828]: I1210 20:18:21.540762 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" exitCode=0 Dec 10 20:18:21 crc kubenswrapper[4828]: I1210 20:18:21.540850 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7"} Dec 10 20:18:21 crc kubenswrapper[4828]: I1210 20:18:21.540909 4828 scope.go:117] "RemoveContainer" containerID="7918ddadd3ea00c849d997afdc8c99d59b1ca27ab5e5de92d512e606ec4f36b8" Dec 10 20:18:21 crc kubenswrapper[4828]: I1210 20:18:21.541648 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:18:21 crc kubenswrapper[4828]: E1210 20:18:21.542064 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:18:36 crc kubenswrapper[4828]: I1210 20:18:36.790074 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:18:36 crc kubenswrapper[4828]: E1210 20:18:36.791228 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:18:51 crc kubenswrapper[4828]: I1210 20:18:51.789275 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:18:51 crc kubenswrapper[4828]: E1210 20:18:51.790440 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:19:06 crc kubenswrapper[4828]: I1210 20:19:06.789741 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:19:06 crc kubenswrapper[4828]: E1210 20:19:06.790681 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:19:21 crc kubenswrapper[4828]: I1210 20:19:21.789549 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:19:21 crc kubenswrapper[4828]: E1210 20:19:21.790618 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:19:32 crc kubenswrapper[4828]: I1210 20:19:32.789014 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:19:32 crc kubenswrapper[4828]: E1210 20:19:32.790057 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:19:44 crc kubenswrapper[4828]: I1210 20:19:44.407469 4828 scope.go:117] "RemoveContainer" containerID="bce6b57156172c42bb1c6ac26dbd0dd1ad0890a766cc9c38a9c3f77b16254160" Dec 10 20:19:44 crc kubenswrapper[4828]: I1210 20:19:44.444761 4828 scope.go:117] "RemoveContainer" containerID="9fc13402bdda2f592cd72099b269c99ec65f4b22c2cee6fbd040310fc276f479" Dec 10 20:19:44 crc kubenswrapper[4828]: I1210 20:19:44.789028 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:19:44 crc kubenswrapper[4828]: E1210 20:19:44.790086 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:19:56 crc kubenswrapper[4828]: I1210 20:19:56.788579 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:19:56 crc kubenswrapper[4828]: E1210 20:19:56.789422 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:20:08 crc kubenswrapper[4828]: I1210 20:20:08.790222 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:20:08 crc kubenswrapper[4828]: E1210 20:20:08.791041 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:20:19 crc kubenswrapper[4828]: I1210 20:20:19.805632 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:20:19 crc kubenswrapper[4828]: E1210 20:20:19.806940 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:20:30 crc kubenswrapper[4828]: I1210 20:20:30.790036 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:20:30 crc kubenswrapper[4828]: E1210 20:20:30.791586 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:20:44 crc kubenswrapper[4828]: I1210 20:20:44.520118 4828 scope.go:117] "RemoveContainer" containerID="539eab4fa21022f1794d7d47c4a9c99691531f43cac9af3039c1c977a90cf837" Dec 10 20:20:45 crc kubenswrapper[4828]: I1210 20:20:45.792706 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:20:45 crc kubenswrapper[4828]: E1210 20:20:45.792973 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:21:00 crc kubenswrapper[4828]: I1210 20:21:00.789235 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:21:00 crc kubenswrapper[4828]: E1210 20:21:00.790003 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:21:14 crc kubenswrapper[4828]: I1210 20:21:14.789684 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:21:14 crc kubenswrapper[4828]: E1210 20:21:14.790752 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:21:28 crc kubenswrapper[4828]: I1210 20:21:28.789507 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:21:28 crc kubenswrapper[4828]: E1210 20:21:28.790235 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:21:40 crc kubenswrapper[4828]: I1210 20:21:40.790068 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:21:40 crc kubenswrapper[4828]: E1210 20:21:40.791054 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:21:54 crc kubenswrapper[4828]: I1210 20:21:54.790136 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:21:54 crc kubenswrapper[4828]: E1210 20:21:54.791480 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:22:07 crc kubenswrapper[4828]: I1210 20:22:07.789467 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:22:07 crc kubenswrapper[4828]: E1210 20:22:07.790690 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:22:22 crc kubenswrapper[4828]: I1210 20:22:22.788637 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:22:22 crc kubenswrapper[4828]: E1210 20:22:22.789452 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:22:36 crc kubenswrapper[4828]: I1210 20:22:36.789893 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:22:36 crc kubenswrapper[4828]: E1210 20:22:36.791449 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:22:51 crc kubenswrapper[4828]: I1210 20:22:51.789151 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:22:51 crc kubenswrapper[4828]: E1210 20:22:51.790097 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:23:06 crc kubenswrapper[4828]: I1210 20:23:06.789708 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:23:06 crc kubenswrapper[4828]: E1210 20:23:06.790644 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:23:21 crc kubenswrapper[4828]: I1210 20:23:21.789171 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:23:22 crc kubenswrapper[4828]: I1210 20:23:22.135986 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"3886f27f4fecd3d3e16c0bed1ffc46840f19c5699ccc27978965f246e4eaad13"} Dec 10 20:24:03 crc kubenswrapper[4828]: I1210 20:24:03.812016 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h8kmv"] Dec 10 20:24:03 crc kubenswrapper[4828]: E1210 20:24:03.813982 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60469732-5c75-4ff5-b0e4-752b73a0af26" containerName="registry-server" Dec 10 20:24:03 crc kubenswrapper[4828]: I1210 20:24:03.813998 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="60469732-5c75-4ff5-b0e4-752b73a0af26" containerName="registry-server" Dec 10 20:24:03 crc kubenswrapper[4828]: E1210 20:24:03.814008 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60469732-5c75-4ff5-b0e4-752b73a0af26" containerName="extract-utilities" Dec 10 20:24:03 crc kubenswrapper[4828]: I1210 20:24:03.814016 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="60469732-5c75-4ff5-b0e4-752b73a0af26" containerName="extract-utilities" Dec 10 20:24:03 crc kubenswrapper[4828]: E1210 20:24:03.814043 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f59f9dc7-df56-4ff1-bf33-0b8c172854f3" containerName="extract-utilities" Dec 10 20:24:03 crc kubenswrapper[4828]: I1210 20:24:03.814049 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="f59f9dc7-df56-4ff1-bf33-0b8c172854f3" containerName="extract-utilities" Dec 10 20:24:03 crc kubenswrapper[4828]: E1210 20:24:03.814062 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60469732-5c75-4ff5-b0e4-752b73a0af26" containerName="extract-content" Dec 10 20:24:03 crc kubenswrapper[4828]: I1210 20:24:03.814068 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="60469732-5c75-4ff5-b0e4-752b73a0af26" containerName="extract-content" Dec 10 20:24:03 crc kubenswrapper[4828]: E1210 20:24:03.814084 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f59f9dc7-df56-4ff1-bf33-0b8c172854f3" containerName="registry-server" Dec 10 20:24:03 crc kubenswrapper[4828]: I1210 20:24:03.814090 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="f59f9dc7-df56-4ff1-bf33-0b8c172854f3" containerName="registry-server" Dec 10 20:24:03 crc kubenswrapper[4828]: E1210 20:24:03.814109 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f59f9dc7-df56-4ff1-bf33-0b8c172854f3" containerName="extract-content" Dec 10 20:24:03 crc kubenswrapper[4828]: I1210 20:24:03.814114 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="f59f9dc7-df56-4ff1-bf33-0b8c172854f3" containerName="extract-content" Dec 10 20:24:03 crc kubenswrapper[4828]: I1210 20:24:03.814363 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="60469732-5c75-4ff5-b0e4-752b73a0af26" containerName="registry-server" Dec 10 20:24:03 crc kubenswrapper[4828]: I1210 20:24:03.814391 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="f59f9dc7-df56-4ff1-bf33-0b8c172854f3" containerName="registry-server" Dec 10 20:24:03 crc kubenswrapper[4828]: I1210 20:24:03.816322 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h8kmv" Dec 10 20:24:03 crc kubenswrapper[4828]: I1210 20:24:03.823699 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h8kmv"] Dec 10 20:24:03 crc kubenswrapper[4828]: I1210 20:24:03.951160 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44ce54a4-be94-4074-80cb-5c79451d4cb1-catalog-content\") pod \"redhat-operators-h8kmv\" (UID: \"44ce54a4-be94-4074-80cb-5c79451d4cb1\") " pod="openshift-marketplace/redhat-operators-h8kmv" Dec 10 20:24:03 crc kubenswrapper[4828]: I1210 20:24:03.951447 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44ce54a4-be94-4074-80cb-5c79451d4cb1-utilities\") pod \"redhat-operators-h8kmv\" (UID: \"44ce54a4-be94-4074-80cb-5c79451d4cb1\") " pod="openshift-marketplace/redhat-operators-h8kmv" Dec 10 20:24:03 crc kubenswrapper[4828]: I1210 20:24:03.951511 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jflg\" (UniqueName: \"kubernetes.io/projected/44ce54a4-be94-4074-80cb-5c79451d4cb1-kube-api-access-8jflg\") pod \"redhat-operators-h8kmv\" (UID: \"44ce54a4-be94-4074-80cb-5c79451d4cb1\") " pod="openshift-marketplace/redhat-operators-h8kmv" Dec 10 20:24:04 crc kubenswrapper[4828]: I1210 20:24:04.053491 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44ce54a4-be94-4074-80cb-5c79451d4cb1-catalog-content\") pod \"redhat-operators-h8kmv\" (UID: \"44ce54a4-be94-4074-80cb-5c79451d4cb1\") " pod="openshift-marketplace/redhat-operators-h8kmv" Dec 10 20:24:04 crc kubenswrapper[4828]: I1210 20:24:04.053641 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44ce54a4-be94-4074-80cb-5c79451d4cb1-utilities\") pod \"redhat-operators-h8kmv\" (UID: \"44ce54a4-be94-4074-80cb-5c79451d4cb1\") " pod="openshift-marketplace/redhat-operators-h8kmv" Dec 10 20:24:04 crc kubenswrapper[4828]: I1210 20:24:04.053712 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jflg\" (UniqueName: \"kubernetes.io/projected/44ce54a4-be94-4074-80cb-5c79451d4cb1-kube-api-access-8jflg\") pod \"redhat-operators-h8kmv\" (UID: \"44ce54a4-be94-4074-80cb-5c79451d4cb1\") " pod="openshift-marketplace/redhat-operators-h8kmv" Dec 10 20:24:04 crc kubenswrapper[4828]: I1210 20:24:04.054178 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44ce54a4-be94-4074-80cb-5c79451d4cb1-catalog-content\") pod \"redhat-operators-h8kmv\" (UID: \"44ce54a4-be94-4074-80cb-5c79451d4cb1\") " pod="openshift-marketplace/redhat-operators-h8kmv" Dec 10 20:24:04 crc kubenswrapper[4828]: I1210 20:24:04.054349 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44ce54a4-be94-4074-80cb-5c79451d4cb1-utilities\") pod \"redhat-operators-h8kmv\" (UID: \"44ce54a4-be94-4074-80cb-5c79451d4cb1\") " pod="openshift-marketplace/redhat-operators-h8kmv" Dec 10 20:24:04 crc kubenswrapper[4828]: I1210 20:24:04.076621 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jflg\" (UniqueName: \"kubernetes.io/projected/44ce54a4-be94-4074-80cb-5c79451d4cb1-kube-api-access-8jflg\") pod \"redhat-operators-h8kmv\" (UID: \"44ce54a4-be94-4074-80cb-5c79451d4cb1\") " pod="openshift-marketplace/redhat-operators-h8kmv" Dec 10 20:24:04 crc kubenswrapper[4828]: I1210 20:24:04.140267 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h8kmv" Dec 10 20:24:04 crc kubenswrapper[4828]: I1210 20:24:04.631819 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h8kmv"] Dec 10 20:24:04 crc kubenswrapper[4828]: I1210 20:24:04.669536 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8kmv" event={"ID":"44ce54a4-be94-4074-80cb-5c79451d4cb1","Type":"ContainerStarted","Data":"62360e8afb1699080874dc558a1a9440318bcf8c8a4d9f0d8a5d46536a41e310"} Dec 10 20:24:05 crc kubenswrapper[4828]: I1210 20:24:05.681228 4828 generic.go:334] "Generic (PLEG): container finished" podID="44ce54a4-be94-4074-80cb-5c79451d4cb1" containerID="5a2ee18024930cd3abe8131ec39d46f8707f48532c55b5b17c3df74ff81e402c" exitCode=0 Dec 10 20:24:05 crc kubenswrapper[4828]: I1210 20:24:05.681286 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8kmv" event={"ID":"44ce54a4-be94-4074-80cb-5c79451d4cb1","Type":"ContainerDied","Data":"5a2ee18024930cd3abe8131ec39d46f8707f48532c55b5b17c3df74ff81e402c"} Dec 10 20:24:05 crc kubenswrapper[4828]: I1210 20:24:05.683378 4828 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 20:24:06 crc kubenswrapper[4828]: I1210 20:24:06.696254 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8kmv" event={"ID":"44ce54a4-be94-4074-80cb-5c79451d4cb1","Type":"ContainerStarted","Data":"96a894229469476c9caca51309b5d508ca448234429a4a9442cc0a72ba0fcc1c"} Dec 10 20:24:09 crc kubenswrapper[4828]: I1210 20:24:09.740899 4828 generic.go:334] "Generic (PLEG): container finished" podID="44ce54a4-be94-4074-80cb-5c79451d4cb1" containerID="96a894229469476c9caca51309b5d508ca448234429a4a9442cc0a72ba0fcc1c" exitCode=0 Dec 10 20:24:09 crc kubenswrapper[4828]: I1210 20:24:09.741002 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8kmv" event={"ID":"44ce54a4-be94-4074-80cb-5c79451d4cb1","Type":"ContainerDied","Data":"96a894229469476c9caca51309b5d508ca448234429a4a9442cc0a72ba0fcc1c"} Dec 10 20:24:10 crc kubenswrapper[4828]: I1210 20:24:10.758930 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8kmv" event={"ID":"44ce54a4-be94-4074-80cb-5c79451d4cb1","Type":"ContainerStarted","Data":"891b1cd1a12166b536efa19663b73cf473adb3370977dd16d5d99f1a6b561d1d"} Dec 10 20:24:10 crc kubenswrapper[4828]: I1210 20:24:10.797193 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h8kmv" podStartSLOduration=3.069658919 podStartE2EDuration="7.79716478s" podCreationTimestamp="2025-12-10 20:24:03 +0000 UTC" firstStartedPulling="2025-12-10 20:24:05.683125089 +0000 UTC m=+5326.193736084" lastFinishedPulling="2025-12-10 20:24:10.4106309 +0000 UTC m=+5330.921241945" observedRunningTime="2025-12-10 20:24:10.782536131 +0000 UTC m=+5331.293147146" watchObservedRunningTime="2025-12-10 20:24:10.79716478 +0000 UTC m=+5331.307775795" Dec 10 20:24:14 crc kubenswrapper[4828]: I1210 20:24:14.141041 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h8kmv" Dec 10 20:24:14 crc kubenswrapper[4828]: I1210 20:24:14.141561 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h8kmv" Dec 10 20:24:15 crc kubenswrapper[4828]: I1210 20:24:15.192118 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-h8kmv" podUID="44ce54a4-be94-4074-80cb-5c79451d4cb1" containerName="registry-server" probeResult="failure" output=< Dec 10 20:24:15 crc kubenswrapper[4828]: timeout: failed to connect service ":50051" within 1s Dec 10 20:24:15 crc kubenswrapper[4828]: > Dec 10 20:24:24 crc kubenswrapper[4828]: I1210 20:24:24.189106 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h8kmv" Dec 10 20:24:24 crc kubenswrapper[4828]: I1210 20:24:24.242770 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h8kmv" Dec 10 20:24:24 crc kubenswrapper[4828]: I1210 20:24:24.437046 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h8kmv"] Dec 10 20:24:25 crc kubenswrapper[4828]: I1210 20:24:25.953755 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h8kmv" podUID="44ce54a4-be94-4074-80cb-5c79451d4cb1" containerName="registry-server" containerID="cri-o://891b1cd1a12166b536efa19663b73cf473adb3370977dd16d5d99f1a6b561d1d" gracePeriod=2 Dec 10 20:24:26 crc kubenswrapper[4828]: I1210 20:24:26.498711 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h8kmv" Dec 10 20:24:26 crc kubenswrapper[4828]: I1210 20:24:26.596608 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44ce54a4-be94-4074-80cb-5c79451d4cb1-utilities\") pod \"44ce54a4-be94-4074-80cb-5c79451d4cb1\" (UID: \"44ce54a4-be94-4074-80cb-5c79451d4cb1\") " Dec 10 20:24:26 crc kubenswrapper[4828]: I1210 20:24:26.596760 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jflg\" (UniqueName: \"kubernetes.io/projected/44ce54a4-be94-4074-80cb-5c79451d4cb1-kube-api-access-8jflg\") pod \"44ce54a4-be94-4074-80cb-5c79451d4cb1\" (UID: \"44ce54a4-be94-4074-80cb-5c79451d4cb1\") " Dec 10 20:24:26 crc kubenswrapper[4828]: I1210 20:24:26.596968 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44ce54a4-be94-4074-80cb-5c79451d4cb1-catalog-content\") pod \"44ce54a4-be94-4074-80cb-5c79451d4cb1\" (UID: \"44ce54a4-be94-4074-80cb-5c79451d4cb1\") " Dec 10 20:24:26 crc kubenswrapper[4828]: I1210 20:24:26.597702 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44ce54a4-be94-4074-80cb-5c79451d4cb1-utilities" (OuterVolumeSpecName: "utilities") pod "44ce54a4-be94-4074-80cb-5c79451d4cb1" (UID: "44ce54a4-be94-4074-80cb-5c79451d4cb1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:24:26 crc kubenswrapper[4828]: I1210 20:24:26.606233 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44ce54a4-be94-4074-80cb-5c79451d4cb1-kube-api-access-8jflg" (OuterVolumeSpecName: "kube-api-access-8jflg") pod "44ce54a4-be94-4074-80cb-5c79451d4cb1" (UID: "44ce54a4-be94-4074-80cb-5c79451d4cb1"). InnerVolumeSpecName "kube-api-access-8jflg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:24:26 crc kubenswrapper[4828]: I1210 20:24:26.699649 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44ce54a4-be94-4074-80cb-5c79451d4cb1-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:24:26 crc kubenswrapper[4828]: I1210 20:24:26.699695 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jflg\" (UniqueName: \"kubernetes.io/projected/44ce54a4-be94-4074-80cb-5c79451d4cb1-kube-api-access-8jflg\") on node \"crc\" DevicePath \"\"" Dec 10 20:24:26 crc kubenswrapper[4828]: I1210 20:24:26.728392 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44ce54a4-be94-4074-80cb-5c79451d4cb1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44ce54a4-be94-4074-80cb-5c79451d4cb1" (UID: "44ce54a4-be94-4074-80cb-5c79451d4cb1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:24:26 crc kubenswrapper[4828]: I1210 20:24:26.802103 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44ce54a4-be94-4074-80cb-5c79451d4cb1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:24:26 crc kubenswrapper[4828]: I1210 20:24:26.972722 4828 generic.go:334] "Generic (PLEG): container finished" podID="44ce54a4-be94-4074-80cb-5c79451d4cb1" containerID="891b1cd1a12166b536efa19663b73cf473adb3370977dd16d5d99f1a6b561d1d" exitCode=0 Dec 10 20:24:26 crc kubenswrapper[4828]: I1210 20:24:26.972789 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8kmv" event={"ID":"44ce54a4-be94-4074-80cb-5c79451d4cb1","Type":"ContainerDied","Data":"891b1cd1a12166b536efa19663b73cf473adb3370977dd16d5d99f1a6b561d1d"} Dec 10 20:24:26 crc kubenswrapper[4828]: I1210 20:24:26.973077 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h8kmv" event={"ID":"44ce54a4-be94-4074-80cb-5c79451d4cb1","Type":"ContainerDied","Data":"62360e8afb1699080874dc558a1a9440318bcf8c8a4d9f0d8a5d46536a41e310"} Dec 10 20:24:26 crc kubenswrapper[4828]: I1210 20:24:26.973101 4828 scope.go:117] "RemoveContainer" containerID="891b1cd1a12166b536efa19663b73cf473adb3370977dd16d5d99f1a6b561d1d" Dec 10 20:24:26 crc kubenswrapper[4828]: I1210 20:24:26.972861 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h8kmv" Dec 10 20:24:27 crc kubenswrapper[4828]: I1210 20:24:27.013567 4828 scope.go:117] "RemoveContainer" containerID="96a894229469476c9caca51309b5d508ca448234429a4a9442cc0a72ba0fcc1c" Dec 10 20:24:27 crc kubenswrapper[4828]: I1210 20:24:27.020899 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h8kmv"] Dec 10 20:24:27 crc kubenswrapper[4828]: I1210 20:24:27.033051 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h8kmv"] Dec 10 20:24:27 crc kubenswrapper[4828]: I1210 20:24:27.044994 4828 scope.go:117] "RemoveContainer" containerID="5a2ee18024930cd3abe8131ec39d46f8707f48532c55b5b17c3df74ff81e402c" Dec 10 20:24:27 crc kubenswrapper[4828]: I1210 20:24:27.096536 4828 scope.go:117] "RemoveContainer" containerID="891b1cd1a12166b536efa19663b73cf473adb3370977dd16d5d99f1a6b561d1d" Dec 10 20:24:27 crc kubenswrapper[4828]: E1210 20:24:27.097023 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"891b1cd1a12166b536efa19663b73cf473adb3370977dd16d5d99f1a6b561d1d\": container with ID starting with 891b1cd1a12166b536efa19663b73cf473adb3370977dd16d5d99f1a6b561d1d not found: ID does not exist" containerID="891b1cd1a12166b536efa19663b73cf473adb3370977dd16d5d99f1a6b561d1d" Dec 10 20:24:27 crc kubenswrapper[4828]: I1210 20:24:27.097134 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"891b1cd1a12166b536efa19663b73cf473adb3370977dd16d5d99f1a6b561d1d"} err="failed to get container status \"891b1cd1a12166b536efa19663b73cf473adb3370977dd16d5d99f1a6b561d1d\": rpc error: code = NotFound desc = could not find container \"891b1cd1a12166b536efa19663b73cf473adb3370977dd16d5d99f1a6b561d1d\": container with ID starting with 891b1cd1a12166b536efa19663b73cf473adb3370977dd16d5d99f1a6b561d1d not found: ID does not exist" Dec 10 20:24:27 crc kubenswrapper[4828]: I1210 20:24:27.097254 4828 scope.go:117] "RemoveContainer" containerID="96a894229469476c9caca51309b5d508ca448234429a4a9442cc0a72ba0fcc1c" Dec 10 20:24:27 crc kubenswrapper[4828]: E1210 20:24:27.097580 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96a894229469476c9caca51309b5d508ca448234429a4a9442cc0a72ba0fcc1c\": container with ID starting with 96a894229469476c9caca51309b5d508ca448234429a4a9442cc0a72ba0fcc1c not found: ID does not exist" containerID="96a894229469476c9caca51309b5d508ca448234429a4a9442cc0a72ba0fcc1c" Dec 10 20:24:27 crc kubenswrapper[4828]: I1210 20:24:27.097668 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96a894229469476c9caca51309b5d508ca448234429a4a9442cc0a72ba0fcc1c"} err="failed to get container status \"96a894229469476c9caca51309b5d508ca448234429a4a9442cc0a72ba0fcc1c\": rpc error: code = NotFound desc = could not find container \"96a894229469476c9caca51309b5d508ca448234429a4a9442cc0a72ba0fcc1c\": container with ID starting with 96a894229469476c9caca51309b5d508ca448234429a4a9442cc0a72ba0fcc1c not found: ID does not exist" Dec 10 20:24:27 crc kubenswrapper[4828]: I1210 20:24:27.097745 4828 scope.go:117] "RemoveContainer" containerID="5a2ee18024930cd3abe8131ec39d46f8707f48532c55b5b17c3df74ff81e402c" Dec 10 20:24:27 crc kubenswrapper[4828]: E1210 20:24:27.098207 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a2ee18024930cd3abe8131ec39d46f8707f48532c55b5b17c3df74ff81e402c\": container with ID starting with 5a2ee18024930cd3abe8131ec39d46f8707f48532c55b5b17c3df74ff81e402c not found: ID does not exist" containerID="5a2ee18024930cd3abe8131ec39d46f8707f48532c55b5b17c3df74ff81e402c" Dec 10 20:24:27 crc kubenswrapper[4828]: I1210 20:24:27.098259 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a2ee18024930cd3abe8131ec39d46f8707f48532c55b5b17c3df74ff81e402c"} err="failed to get container status \"5a2ee18024930cd3abe8131ec39d46f8707f48532c55b5b17c3df74ff81e402c\": rpc error: code = NotFound desc = could not find container \"5a2ee18024930cd3abe8131ec39d46f8707f48532c55b5b17c3df74ff81e402c\": container with ID starting with 5a2ee18024930cd3abe8131ec39d46f8707f48532c55b5b17c3df74ff81e402c not found: ID does not exist" Dec 10 20:24:27 crc kubenswrapper[4828]: I1210 20:24:27.809315 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44ce54a4-be94-4074-80cb-5c79451d4cb1" path="/var/lib/kubelet/pods/44ce54a4-be94-4074-80cb-5c79451d4cb1/volumes" Dec 10 20:25:51 crc kubenswrapper[4828]: I1210 20:25:51.230550 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:25:51 crc kubenswrapper[4828]: I1210 20:25:51.231600 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:26:18 crc kubenswrapper[4828]: I1210 20:26:18.867466 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8vpqj"] Dec 10 20:26:18 crc kubenswrapper[4828]: E1210 20:26:18.868679 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44ce54a4-be94-4074-80cb-5c79451d4cb1" containerName="extract-utilities" Dec 10 20:26:18 crc kubenswrapper[4828]: I1210 20:26:18.868697 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="44ce54a4-be94-4074-80cb-5c79451d4cb1" containerName="extract-utilities" Dec 10 20:26:18 crc kubenswrapper[4828]: E1210 20:26:18.868729 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44ce54a4-be94-4074-80cb-5c79451d4cb1" containerName="extract-content" Dec 10 20:26:18 crc kubenswrapper[4828]: I1210 20:26:18.868738 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="44ce54a4-be94-4074-80cb-5c79451d4cb1" containerName="extract-content" Dec 10 20:26:18 crc kubenswrapper[4828]: E1210 20:26:18.868769 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44ce54a4-be94-4074-80cb-5c79451d4cb1" containerName="registry-server" Dec 10 20:26:18 crc kubenswrapper[4828]: I1210 20:26:18.868779 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="44ce54a4-be94-4074-80cb-5c79451d4cb1" containerName="registry-server" Dec 10 20:26:18 crc kubenswrapper[4828]: I1210 20:26:18.869151 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="44ce54a4-be94-4074-80cb-5c79451d4cb1" containerName="registry-server" Dec 10 20:26:18 crc kubenswrapper[4828]: I1210 20:26:18.871292 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vpqj" Dec 10 20:26:18 crc kubenswrapper[4828]: I1210 20:26:18.886464 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8vpqj"] Dec 10 20:26:18 crc kubenswrapper[4828]: I1210 20:26:18.964021 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9350ad6d-b172-4eae-93f4-c1e5d92c1990-catalog-content\") pod \"certified-operators-8vpqj\" (UID: \"9350ad6d-b172-4eae-93f4-c1e5d92c1990\") " pod="openshift-marketplace/certified-operators-8vpqj" Dec 10 20:26:18 crc kubenswrapper[4828]: I1210 20:26:18.964272 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfwwq\" (UniqueName: \"kubernetes.io/projected/9350ad6d-b172-4eae-93f4-c1e5d92c1990-kube-api-access-vfwwq\") pod \"certified-operators-8vpqj\" (UID: \"9350ad6d-b172-4eae-93f4-c1e5d92c1990\") " pod="openshift-marketplace/certified-operators-8vpqj" Dec 10 20:26:18 crc kubenswrapper[4828]: I1210 20:26:18.964433 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9350ad6d-b172-4eae-93f4-c1e5d92c1990-utilities\") pod \"certified-operators-8vpqj\" (UID: \"9350ad6d-b172-4eae-93f4-c1e5d92c1990\") " pod="openshift-marketplace/certified-operators-8vpqj" Dec 10 20:26:19 crc kubenswrapper[4828]: I1210 20:26:19.067766 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9350ad6d-b172-4eae-93f4-c1e5d92c1990-catalog-content\") pod \"certified-operators-8vpqj\" (UID: \"9350ad6d-b172-4eae-93f4-c1e5d92c1990\") " pod="openshift-marketplace/certified-operators-8vpqj" Dec 10 20:26:19 crc kubenswrapper[4828]: I1210 20:26:19.067874 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfwwq\" (UniqueName: \"kubernetes.io/projected/9350ad6d-b172-4eae-93f4-c1e5d92c1990-kube-api-access-vfwwq\") pod \"certified-operators-8vpqj\" (UID: \"9350ad6d-b172-4eae-93f4-c1e5d92c1990\") " pod="openshift-marketplace/certified-operators-8vpqj" Dec 10 20:26:19 crc kubenswrapper[4828]: I1210 20:26:19.068137 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9350ad6d-b172-4eae-93f4-c1e5d92c1990-utilities\") pod \"certified-operators-8vpqj\" (UID: \"9350ad6d-b172-4eae-93f4-c1e5d92c1990\") " pod="openshift-marketplace/certified-operators-8vpqj" Dec 10 20:26:19 crc kubenswrapper[4828]: I1210 20:26:19.068592 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9350ad6d-b172-4eae-93f4-c1e5d92c1990-utilities\") pod \"certified-operators-8vpqj\" (UID: \"9350ad6d-b172-4eae-93f4-c1e5d92c1990\") " pod="openshift-marketplace/certified-operators-8vpqj" Dec 10 20:26:19 crc kubenswrapper[4828]: I1210 20:26:19.068733 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9350ad6d-b172-4eae-93f4-c1e5d92c1990-catalog-content\") pod \"certified-operators-8vpqj\" (UID: \"9350ad6d-b172-4eae-93f4-c1e5d92c1990\") " pod="openshift-marketplace/certified-operators-8vpqj" Dec 10 20:26:19 crc kubenswrapper[4828]: I1210 20:26:19.095182 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfwwq\" (UniqueName: \"kubernetes.io/projected/9350ad6d-b172-4eae-93f4-c1e5d92c1990-kube-api-access-vfwwq\") pod \"certified-operators-8vpqj\" (UID: \"9350ad6d-b172-4eae-93f4-c1e5d92c1990\") " pod="openshift-marketplace/certified-operators-8vpqj" Dec 10 20:26:19 crc kubenswrapper[4828]: I1210 20:26:19.190441 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vpqj" Dec 10 20:26:19 crc kubenswrapper[4828]: I1210 20:26:19.664119 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8vpqj"] Dec 10 20:26:20 crc kubenswrapper[4828]: I1210 20:26:20.519192 4828 generic.go:334] "Generic (PLEG): container finished" podID="9350ad6d-b172-4eae-93f4-c1e5d92c1990" containerID="138d0454593c71ebcdc6f6e8d09ad110b4f871d0a7a565413795aa508e106a22" exitCode=0 Dec 10 20:26:20 crc kubenswrapper[4828]: I1210 20:26:20.519721 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vpqj" event={"ID":"9350ad6d-b172-4eae-93f4-c1e5d92c1990","Type":"ContainerDied","Data":"138d0454593c71ebcdc6f6e8d09ad110b4f871d0a7a565413795aa508e106a22"} Dec 10 20:26:20 crc kubenswrapper[4828]: I1210 20:26:20.519745 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vpqj" event={"ID":"9350ad6d-b172-4eae-93f4-c1e5d92c1990","Type":"ContainerStarted","Data":"26a06e8fd550576ff685ac9dc5396c13196d2600c0dc5f5b0070eca598af9c17"} Dec 10 20:26:21 crc kubenswrapper[4828]: I1210 20:26:21.230406 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:26:21 crc kubenswrapper[4828]: I1210 20:26:21.230756 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:26:22 crc kubenswrapper[4828]: I1210 20:26:22.557174 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vpqj" event={"ID":"9350ad6d-b172-4eae-93f4-c1e5d92c1990","Type":"ContainerStarted","Data":"20fff3fc157d99b5edbdb3aba21840d4aac6723422104431e747037798a29e0a"} Dec 10 20:26:23 crc kubenswrapper[4828]: I1210 20:26:23.570976 4828 generic.go:334] "Generic (PLEG): container finished" podID="9350ad6d-b172-4eae-93f4-c1e5d92c1990" containerID="20fff3fc157d99b5edbdb3aba21840d4aac6723422104431e747037798a29e0a" exitCode=0 Dec 10 20:26:23 crc kubenswrapper[4828]: I1210 20:26:23.571022 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vpqj" event={"ID":"9350ad6d-b172-4eae-93f4-c1e5d92c1990","Type":"ContainerDied","Data":"20fff3fc157d99b5edbdb3aba21840d4aac6723422104431e747037798a29e0a"} Dec 10 20:26:24 crc kubenswrapper[4828]: I1210 20:26:24.583085 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vpqj" event={"ID":"9350ad6d-b172-4eae-93f4-c1e5d92c1990","Type":"ContainerStarted","Data":"e2e0603bb566f70fbc145f05c5ba02f3bb985068e998d10ba22c82a44f4ce566"} Dec 10 20:26:24 crc kubenswrapper[4828]: I1210 20:26:24.611121 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8vpqj" podStartSLOduration=3.069995038 podStartE2EDuration="6.611098653s" podCreationTimestamp="2025-12-10 20:26:18 +0000 UTC" firstStartedPulling="2025-12-10 20:26:20.521241356 +0000 UTC m=+5461.031852361" lastFinishedPulling="2025-12-10 20:26:24.062344971 +0000 UTC m=+5464.572955976" observedRunningTime="2025-12-10 20:26:24.597852339 +0000 UTC m=+5465.108463364" watchObservedRunningTime="2025-12-10 20:26:24.611098653 +0000 UTC m=+5465.121709658" Dec 10 20:26:29 crc kubenswrapper[4828]: I1210 20:26:29.191164 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8vpqj" Dec 10 20:26:29 crc kubenswrapper[4828]: I1210 20:26:29.192394 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8vpqj" Dec 10 20:26:30 crc kubenswrapper[4828]: I1210 20:26:30.255724 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-8vpqj" podUID="9350ad6d-b172-4eae-93f4-c1e5d92c1990" containerName="registry-server" probeResult="failure" output=< Dec 10 20:26:30 crc kubenswrapper[4828]: timeout: failed to connect service ":50051" within 1s Dec 10 20:26:30 crc kubenswrapper[4828]: > Dec 10 20:26:39 crc kubenswrapper[4828]: I1210 20:26:39.260960 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8vpqj" Dec 10 20:26:39 crc kubenswrapper[4828]: I1210 20:26:39.335547 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8vpqj" Dec 10 20:26:39 crc kubenswrapper[4828]: I1210 20:26:39.510779 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8vpqj"] Dec 10 20:26:40 crc kubenswrapper[4828]: I1210 20:26:40.760159 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8vpqj" podUID="9350ad6d-b172-4eae-93f4-c1e5d92c1990" containerName="registry-server" containerID="cri-o://e2e0603bb566f70fbc145f05c5ba02f3bb985068e998d10ba22c82a44f4ce566" gracePeriod=2 Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.272891 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vpqj" Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.315222 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfwwq\" (UniqueName: \"kubernetes.io/projected/9350ad6d-b172-4eae-93f4-c1e5d92c1990-kube-api-access-vfwwq\") pod \"9350ad6d-b172-4eae-93f4-c1e5d92c1990\" (UID: \"9350ad6d-b172-4eae-93f4-c1e5d92c1990\") " Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.315287 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9350ad6d-b172-4eae-93f4-c1e5d92c1990-catalog-content\") pod \"9350ad6d-b172-4eae-93f4-c1e5d92c1990\" (UID: \"9350ad6d-b172-4eae-93f4-c1e5d92c1990\") " Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.315784 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9350ad6d-b172-4eae-93f4-c1e5d92c1990-utilities\") pod \"9350ad6d-b172-4eae-93f4-c1e5d92c1990\" (UID: \"9350ad6d-b172-4eae-93f4-c1e5d92c1990\") " Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.317116 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9350ad6d-b172-4eae-93f4-c1e5d92c1990-utilities" (OuterVolumeSpecName: "utilities") pod "9350ad6d-b172-4eae-93f4-c1e5d92c1990" (UID: "9350ad6d-b172-4eae-93f4-c1e5d92c1990"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.323415 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9350ad6d-b172-4eae-93f4-c1e5d92c1990-kube-api-access-vfwwq" (OuterVolumeSpecName: "kube-api-access-vfwwq") pod "9350ad6d-b172-4eae-93f4-c1e5d92c1990" (UID: "9350ad6d-b172-4eae-93f4-c1e5d92c1990"). InnerVolumeSpecName "kube-api-access-vfwwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.372498 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9350ad6d-b172-4eae-93f4-c1e5d92c1990-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9350ad6d-b172-4eae-93f4-c1e5d92c1990" (UID: "9350ad6d-b172-4eae-93f4-c1e5d92c1990"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.419139 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfwwq\" (UniqueName: \"kubernetes.io/projected/9350ad6d-b172-4eae-93f4-c1e5d92c1990-kube-api-access-vfwwq\") on node \"crc\" DevicePath \"\"" Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.419526 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9350ad6d-b172-4eae-93f4-c1e5d92c1990-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.419686 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9350ad6d-b172-4eae-93f4-c1e5d92c1990-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.776554 4828 generic.go:334] "Generic (PLEG): container finished" podID="9350ad6d-b172-4eae-93f4-c1e5d92c1990" containerID="e2e0603bb566f70fbc145f05c5ba02f3bb985068e998d10ba22c82a44f4ce566" exitCode=0 Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.776662 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8vpqj" Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.776653 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vpqj" event={"ID":"9350ad6d-b172-4eae-93f4-c1e5d92c1990","Type":"ContainerDied","Data":"e2e0603bb566f70fbc145f05c5ba02f3bb985068e998d10ba22c82a44f4ce566"} Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.777201 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8vpqj" event={"ID":"9350ad6d-b172-4eae-93f4-c1e5d92c1990","Type":"ContainerDied","Data":"26a06e8fd550576ff685ac9dc5396c13196d2600c0dc5f5b0070eca598af9c17"} Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.777243 4828 scope.go:117] "RemoveContainer" containerID="e2e0603bb566f70fbc145f05c5ba02f3bb985068e998d10ba22c82a44f4ce566" Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.814084 4828 scope.go:117] "RemoveContainer" containerID="20fff3fc157d99b5edbdb3aba21840d4aac6723422104431e747037798a29e0a" Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.833096 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8vpqj"] Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.847783 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8vpqj"] Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.851628 4828 scope.go:117] "RemoveContainer" containerID="138d0454593c71ebcdc6f6e8d09ad110b4f871d0a7a565413795aa508e106a22" Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.916401 4828 scope.go:117] "RemoveContainer" containerID="e2e0603bb566f70fbc145f05c5ba02f3bb985068e998d10ba22c82a44f4ce566" Dec 10 20:26:41 crc kubenswrapper[4828]: E1210 20:26:41.917218 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2e0603bb566f70fbc145f05c5ba02f3bb985068e998d10ba22c82a44f4ce566\": container with ID starting with e2e0603bb566f70fbc145f05c5ba02f3bb985068e998d10ba22c82a44f4ce566 not found: ID does not exist" containerID="e2e0603bb566f70fbc145f05c5ba02f3bb985068e998d10ba22c82a44f4ce566" Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.917278 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2e0603bb566f70fbc145f05c5ba02f3bb985068e998d10ba22c82a44f4ce566"} err="failed to get container status \"e2e0603bb566f70fbc145f05c5ba02f3bb985068e998d10ba22c82a44f4ce566\": rpc error: code = NotFound desc = could not find container \"e2e0603bb566f70fbc145f05c5ba02f3bb985068e998d10ba22c82a44f4ce566\": container with ID starting with e2e0603bb566f70fbc145f05c5ba02f3bb985068e998d10ba22c82a44f4ce566 not found: ID does not exist" Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.917317 4828 scope.go:117] "RemoveContainer" containerID="20fff3fc157d99b5edbdb3aba21840d4aac6723422104431e747037798a29e0a" Dec 10 20:26:41 crc kubenswrapper[4828]: E1210 20:26:41.917792 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20fff3fc157d99b5edbdb3aba21840d4aac6723422104431e747037798a29e0a\": container with ID starting with 20fff3fc157d99b5edbdb3aba21840d4aac6723422104431e747037798a29e0a not found: ID does not exist" containerID="20fff3fc157d99b5edbdb3aba21840d4aac6723422104431e747037798a29e0a" Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.918033 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20fff3fc157d99b5edbdb3aba21840d4aac6723422104431e747037798a29e0a"} err="failed to get container status \"20fff3fc157d99b5edbdb3aba21840d4aac6723422104431e747037798a29e0a\": rpc error: code = NotFound desc = could not find container \"20fff3fc157d99b5edbdb3aba21840d4aac6723422104431e747037798a29e0a\": container with ID starting with 20fff3fc157d99b5edbdb3aba21840d4aac6723422104431e747037798a29e0a not found: ID does not exist" Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.918050 4828 scope.go:117] "RemoveContainer" containerID="138d0454593c71ebcdc6f6e8d09ad110b4f871d0a7a565413795aa508e106a22" Dec 10 20:26:41 crc kubenswrapper[4828]: E1210 20:26:41.918459 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"138d0454593c71ebcdc6f6e8d09ad110b4f871d0a7a565413795aa508e106a22\": container with ID starting with 138d0454593c71ebcdc6f6e8d09ad110b4f871d0a7a565413795aa508e106a22 not found: ID does not exist" containerID="138d0454593c71ebcdc6f6e8d09ad110b4f871d0a7a565413795aa508e106a22" Dec 10 20:26:41 crc kubenswrapper[4828]: I1210 20:26:41.918530 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"138d0454593c71ebcdc6f6e8d09ad110b4f871d0a7a565413795aa508e106a22"} err="failed to get container status \"138d0454593c71ebcdc6f6e8d09ad110b4f871d0a7a565413795aa508e106a22\": rpc error: code = NotFound desc = could not find container \"138d0454593c71ebcdc6f6e8d09ad110b4f871d0a7a565413795aa508e106a22\": container with ID starting with 138d0454593c71ebcdc6f6e8d09ad110b4f871d0a7a565413795aa508e106a22 not found: ID does not exist" Dec 10 20:26:43 crc kubenswrapper[4828]: I1210 20:26:43.813364 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9350ad6d-b172-4eae-93f4-c1e5d92c1990" path="/var/lib/kubelet/pods/9350ad6d-b172-4eae-93f4-c1e5d92c1990/volumes" Dec 10 20:26:51 crc kubenswrapper[4828]: I1210 20:26:51.230667 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:26:51 crc kubenswrapper[4828]: I1210 20:26:51.231315 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:26:51 crc kubenswrapper[4828]: I1210 20:26:51.231363 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 20:26:51 crc kubenswrapper[4828]: I1210 20:26:51.232311 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3886f27f4fecd3d3e16c0bed1ffc46840f19c5699ccc27978965f246e4eaad13"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 20:26:51 crc kubenswrapper[4828]: I1210 20:26:51.232369 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://3886f27f4fecd3d3e16c0bed1ffc46840f19c5699ccc27978965f246e4eaad13" gracePeriod=600 Dec 10 20:26:51 crc kubenswrapper[4828]: I1210 20:26:51.909055 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="3886f27f4fecd3d3e16c0bed1ffc46840f19c5699ccc27978965f246e4eaad13" exitCode=0 Dec 10 20:26:51 crc kubenswrapper[4828]: I1210 20:26:51.909100 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"3886f27f4fecd3d3e16c0bed1ffc46840f19c5699ccc27978965f246e4eaad13"} Dec 10 20:26:51 crc kubenswrapper[4828]: I1210 20:26:51.909608 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2"} Dec 10 20:26:51 crc kubenswrapper[4828]: I1210 20:26:51.909635 4828 scope.go:117] "RemoveContainer" containerID="cd30aafcfb3d42699859194a6213a67ab0701119e09fbedd90dad861adcd92d7" Dec 10 20:27:29 crc kubenswrapper[4828]: I1210 20:27:29.668223 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7tx7z"] Dec 10 20:27:29 crc kubenswrapper[4828]: E1210 20:27:29.669574 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9350ad6d-b172-4eae-93f4-c1e5d92c1990" containerName="extract-utilities" Dec 10 20:27:29 crc kubenswrapper[4828]: I1210 20:27:29.669591 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="9350ad6d-b172-4eae-93f4-c1e5d92c1990" containerName="extract-utilities" Dec 10 20:27:29 crc kubenswrapper[4828]: E1210 20:27:29.669620 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9350ad6d-b172-4eae-93f4-c1e5d92c1990" containerName="extract-content" Dec 10 20:27:29 crc kubenswrapper[4828]: I1210 20:27:29.669629 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="9350ad6d-b172-4eae-93f4-c1e5d92c1990" containerName="extract-content" Dec 10 20:27:29 crc kubenswrapper[4828]: E1210 20:27:29.669650 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9350ad6d-b172-4eae-93f4-c1e5d92c1990" containerName="registry-server" Dec 10 20:27:29 crc kubenswrapper[4828]: I1210 20:27:29.669659 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="9350ad6d-b172-4eae-93f4-c1e5d92c1990" containerName="registry-server" Dec 10 20:27:29 crc kubenswrapper[4828]: I1210 20:27:29.670009 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="9350ad6d-b172-4eae-93f4-c1e5d92c1990" containerName="registry-server" Dec 10 20:27:29 crc kubenswrapper[4828]: I1210 20:27:29.672236 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7tx7z" Dec 10 20:27:29 crc kubenswrapper[4828]: I1210 20:27:29.696832 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7tx7z"] Dec 10 20:27:29 crc kubenswrapper[4828]: I1210 20:27:29.856553 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n9vt\" (UniqueName: \"kubernetes.io/projected/c7848156-6783-4de5-a9cb-07f89c916747-kube-api-access-4n9vt\") pod \"community-operators-7tx7z\" (UID: \"c7848156-6783-4de5-a9cb-07f89c916747\") " pod="openshift-marketplace/community-operators-7tx7z" Dec 10 20:27:29 crc kubenswrapper[4828]: I1210 20:27:29.857006 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7848156-6783-4de5-a9cb-07f89c916747-catalog-content\") pod \"community-operators-7tx7z\" (UID: \"c7848156-6783-4de5-a9cb-07f89c916747\") " pod="openshift-marketplace/community-operators-7tx7z" Dec 10 20:27:29 crc kubenswrapper[4828]: I1210 20:27:29.857105 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7848156-6783-4de5-a9cb-07f89c916747-utilities\") pod \"community-operators-7tx7z\" (UID: \"c7848156-6783-4de5-a9cb-07f89c916747\") " pod="openshift-marketplace/community-operators-7tx7z" Dec 10 20:27:29 crc kubenswrapper[4828]: I1210 20:27:29.959684 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n9vt\" (UniqueName: \"kubernetes.io/projected/c7848156-6783-4de5-a9cb-07f89c916747-kube-api-access-4n9vt\") pod \"community-operators-7tx7z\" (UID: \"c7848156-6783-4de5-a9cb-07f89c916747\") " pod="openshift-marketplace/community-operators-7tx7z" Dec 10 20:27:29 crc kubenswrapper[4828]: I1210 20:27:29.959814 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7848156-6783-4de5-a9cb-07f89c916747-catalog-content\") pod \"community-operators-7tx7z\" (UID: \"c7848156-6783-4de5-a9cb-07f89c916747\") " pod="openshift-marketplace/community-operators-7tx7z" Dec 10 20:27:29 crc kubenswrapper[4828]: I1210 20:27:29.959837 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7848156-6783-4de5-a9cb-07f89c916747-utilities\") pod \"community-operators-7tx7z\" (UID: \"c7848156-6783-4de5-a9cb-07f89c916747\") " pod="openshift-marketplace/community-operators-7tx7z" Dec 10 20:27:29 crc kubenswrapper[4828]: I1210 20:27:29.960546 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7848156-6783-4de5-a9cb-07f89c916747-catalog-content\") pod \"community-operators-7tx7z\" (UID: \"c7848156-6783-4de5-a9cb-07f89c916747\") " pod="openshift-marketplace/community-operators-7tx7z" Dec 10 20:27:29 crc kubenswrapper[4828]: I1210 20:27:29.960628 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7848156-6783-4de5-a9cb-07f89c916747-utilities\") pod \"community-operators-7tx7z\" (UID: \"c7848156-6783-4de5-a9cb-07f89c916747\") " pod="openshift-marketplace/community-operators-7tx7z" Dec 10 20:27:29 crc kubenswrapper[4828]: I1210 20:27:29.981158 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n9vt\" (UniqueName: \"kubernetes.io/projected/c7848156-6783-4de5-a9cb-07f89c916747-kube-api-access-4n9vt\") pod \"community-operators-7tx7z\" (UID: \"c7848156-6783-4de5-a9cb-07f89c916747\") " pod="openshift-marketplace/community-operators-7tx7z" Dec 10 20:27:30 crc kubenswrapper[4828]: I1210 20:27:30.006268 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7tx7z" Dec 10 20:27:30 crc kubenswrapper[4828]: I1210 20:27:30.597963 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7tx7z"] Dec 10 20:27:31 crc kubenswrapper[4828]: I1210 20:27:31.399813 4828 generic.go:334] "Generic (PLEG): container finished" podID="c7848156-6783-4de5-a9cb-07f89c916747" containerID="216d66fdf5c3a310ddaa2c1280c3ec3ff20214e3854b46e6c2675ba1c82ea455" exitCode=0 Dec 10 20:27:31 crc kubenswrapper[4828]: I1210 20:27:31.400244 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7tx7z" event={"ID":"c7848156-6783-4de5-a9cb-07f89c916747","Type":"ContainerDied","Data":"216d66fdf5c3a310ddaa2c1280c3ec3ff20214e3854b46e6c2675ba1c82ea455"} Dec 10 20:27:31 crc kubenswrapper[4828]: I1210 20:27:31.400354 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7tx7z" event={"ID":"c7848156-6783-4de5-a9cb-07f89c916747","Type":"ContainerStarted","Data":"2cd64176c6955900e5a45ec4c2799ac1eefdc75a65e7251f01e192993e8d0671"} Dec 10 20:27:32 crc kubenswrapper[4828]: I1210 20:27:32.424561 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7tx7z" event={"ID":"c7848156-6783-4de5-a9cb-07f89c916747","Type":"ContainerStarted","Data":"423522606e766cfaeac3bce0559152842e13fec946d4aab9f20be3822d088646"} Dec 10 20:27:33 crc kubenswrapper[4828]: I1210 20:27:33.438136 4828 generic.go:334] "Generic (PLEG): container finished" podID="c7848156-6783-4de5-a9cb-07f89c916747" containerID="423522606e766cfaeac3bce0559152842e13fec946d4aab9f20be3822d088646" exitCode=0 Dec 10 20:27:33 crc kubenswrapper[4828]: I1210 20:27:33.438228 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7tx7z" event={"ID":"c7848156-6783-4de5-a9cb-07f89c916747","Type":"ContainerDied","Data":"423522606e766cfaeac3bce0559152842e13fec946d4aab9f20be3822d088646"} Dec 10 20:27:34 crc kubenswrapper[4828]: I1210 20:27:34.452240 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7tx7z" event={"ID":"c7848156-6783-4de5-a9cb-07f89c916747","Type":"ContainerStarted","Data":"097d42aa03a96e9ff52b50b82e30febeb6c3715cbfa456993c905e97f0713278"} Dec 10 20:27:34 crc kubenswrapper[4828]: I1210 20:27:34.482269 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7tx7z" podStartSLOduration=3.034529079 podStartE2EDuration="5.48224401s" podCreationTimestamp="2025-12-10 20:27:29 +0000 UTC" firstStartedPulling="2025-12-10 20:27:31.403938137 +0000 UTC m=+5531.914549142" lastFinishedPulling="2025-12-10 20:27:33.851653068 +0000 UTC m=+5534.362264073" observedRunningTime="2025-12-10 20:27:34.473062754 +0000 UTC m=+5534.983673789" watchObservedRunningTime="2025-12-10 20:27:34.48224401 +0000 UTC m=+5534.992855035" Dec 10 20:27:40 crc kubenswrapper[4828]: I1210 20:27:40.007301 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7tx7z" Dec 10 20:27:40 crc kubenswrapper[4828]: I1210 20:27:40.008083 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7tx7z" Dec 10 20:27:40 crc kubenswrapper[4828]: I1210 20:27:40.074524 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7tx7z" Dec 10 20:27:41 crc kubenswrapper[4828]: I1210 20:27:41.033464 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7tx7z" Dec 10 20:27:41 crc kubenswrapper[4828]: I1210 20:27:41.099072 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7tx7z"] Dec 10 20:27:42 crc kubenswrapper[4828]: I1210 20:27:42.606116 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7tx7z" podUID="c7848156-6783-4de5-a9cb-07f89c916747" containerName="registry-server" containerID="cri-o://097d42aa03a96e9ff52b50b82e30febeb6c3715cbfa456993c905e97f0713278" gracePeriod=2 Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.153207 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7tx7z" Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.255372 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4n9vt\" (UniqueName: \"kubernetes.io/projected/c7848156-6783-4de5-a9cb-07f89c916747-kube-api-access-4n9vt\") pod \"c7848156-6783-4de5-a9cb-07f89c916747\" (UID: \"c7848156-6783-4de5-a9cb-07f89c916747\") " Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.255521 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7848156-6783-4de5-a9cb-07f89c916747-utilities\") pod \"c7848156-6783-4de5-a9cb-07f89c916747\" (UID: \"c7848156-6783-4de5-a9cb-07f89c916747\") " Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.255567 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7848156-6783-4de5-a9cb-07f89c916747-catalog-content\") pod \"c7848156-6783-4de5-a9cb-07f89c916747\" (UID: \"c7848156-6783-4de5-a9cb-07f89c916747\") " Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.257357 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7848156-6783-4de5-a9cb-07f89c916747-utilities" (OuterVolumeSpecName: "utilities") pod "c7848156-6783-4de5-a9cb-07f89c916747" (UID: "c7848156-6783-4de5-a9cb-07f89c916747"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.264999 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7848156-6783-4de5-a9cb-07f89c916747-kube-api-access-4n9vt" (OuterVolumeSpecName: "kube-api-access-4n9vt") pod "c7848156-6783-4de5-a9cb-07f89c916747" (UID: "c7848156-6783-4de5-a9cb-07f89c916747"). InnerVolumeSpecName "kube-api-access-4n9vt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.318564 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7848156-6783-4de5-a9cb-07f89c916747-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c7848156-6783-4de5-a9cb-07f89c916747" (UID: "c7848156-6783-4de5-a9cb-07f89c916747"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.358306 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4n9vt\" (UniqueName: \"kubernetes.io/projected/c7848156-6783-4de5-a9cb-07f89c916747-kube-api-access-4n9vt\") on node \"crc\" DevicePath \"\"" Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.358561 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7848156-6783-4de5-a9cb-07f89c916747-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.358652 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7848156-6783-4de5-a9cb-07f89c916747-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.761439 4828 generic.go:334] "Generic (PLEG): container finished" podID="c7848156-6783-4de5-a9cb-07f89c916747" containerID="097d42aa03a96e9ff52b50b82e30febeb6c3715cbfa456993c905e97f0713278" exitCode=0 Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.761500 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7tx7z" event={"ID":"c7848156-6783-4de5-a9cb-07f89c916747","Type":"ContainerDied","Data":"097d42aa03a96e9ff52b50b82e30febeb6c3715cbfa456993c905e97f0713278"} Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.761531 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7tx7z" Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.761564 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7tx7z" event={"ID":"c7848156-6783-4de5-a9cb-07f89c916747","Type":"ContainerDied","Data":"2cd64176c6955900e5a45ec4c2799ac1eefdc75a65e7251f01e192993e8d0671"} Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.761584 4828 scope.go:117] "RemoveContainer" containerID="097d42aa03a96e9ff52b50b82e30febeb6c3715cbfa456993c905e97f0713278" Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.794097 4828 scope.go:117] "RemoveContainer" containerID="423522606e766cfaeac3bce0559152842e13fec946d4aab9f20be3822d088646" Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.805042 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7tx7z"] Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.814934 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7tx7z"] Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.827623 4828 scope.go:117] "RemoveContainer" containerID="216d66fdf5c3a310ddaa2c1280c3ec3ff20214e3854b46e6c2675ba1c82ea455" Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.875393 4828 scope.go:117] "RemoveContainer" containerID="097d42aa03a96e9ff52b50b82e30febeb6c3715cbfa456993c905e97f0713278" Dec 10 20:27:43 crc kubenswrapper[4828]: E1210 20:27:43.875901 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"097d42aa03a96e9ff52b50b82e30febeb6c3715cbfa456993c905e97f0713278\": container with ID starting with 097d42aa03a96e9ff52b50b82e30febeb6c3715cbfa456993c905e97f0713278 not found: ID does not exist" containerID="097d42aa03a96e9ff52b50b82e30febeb6c3715cbfa456993c905e97f0713278" Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.875945 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"097d42aa03a96e9ff52b50b82e30febeb6c3715cbfa456993c905e97f0713278"} err="failed to get container status \"097d42aa03a96e9ff52b50b82e30febeb6c3715cbfa456993c905e97f0713278\": rpc error: code = NotFound desc = could not find container \"097d42aa03a96e9ff52b50b82e30febeb6c3715cbfa456993c905e97f0713278\": container with ID starting with 097d42aa03a96e9ff52b50b82e30febeb6c3715cbfa456993c905e97f0713278 not found: ID does not exist" Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.875972 4828 scope.go:117] "RemoveContainer" containerID="423522606e766cfaeac3bce0559152842e13fec946d4aab9f20be3822d088646" Dec 10 20:27:43 crc kubenswrapper[4828]: E1210 20:27:43.876517 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"423522606e766cfaeac3bce0559152842e13fec946d4aab9f20be3822d088646\": container with ID starting with 423522606e766cfaeac3bce0559152842e13fec946d4aab9f20be3822d088646 not found: ID does not exist" containerID="423522606e766cfaeac3bce0559152842e13fec946d4aab9f20be3822d088646" Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.876544 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"423522606e766cfaeac3bce0559152842e13fec946d4aab9f20be3822d088646"} err="failed to get container status \"423522606e766cfaeac3bce0559152842e13fec946d4aab9f20be3822d088646\": rpc error: code = NotFound desc = could not find container \"423522606e766cfaeac3bce0559152842e13fec946d4aab9f20be3822d088646\": container with ID starting with 423522606e766cfaeac3bce0559152842e13fec946d4aab9f20be3822d088646 not found: ID does not exist" Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.876561 4828 scope.go:117] "RemoveContainer" containerID="216d66fdf5c3a310ddaa2c1280c3ec3ff20214e3854b46e6c2675ba1c82ea455" Dec 10 20:27:43 crc kubenswrapper[4828]: E1210 20:27:43.876906 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"216d66fdf5c3a310ddaa2c1280c3ec3ff20214e3854b46e6c2675ba1c82ea455\": container with ID starting with 216d66fdf5c3a310ddaa2c1280c3ec3ff20214e3854b46e6c2675ba1c82ea455 not found: ID does not exist" containerID="216d66fdf5c3a310ddaa2c1280c3ec3ff20214e3854b46e6c2675ba1c82ea455" Dec 10 20:27:43 crc kubenswrapper[4828]: I1210 20:27:43.876934 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"216d66fdf5c3a310ddaa2c1280c3ec3ff20214e3854b46e6c2675ba1c82ea455"} err="failed to get container status \"216d66fdf5c3a310ddaa2c1280c3ec3ff20214e3854b46e6c2675ba1c82ea455\": rpc error: code = NotFound desc = could not find container \"216d66fdf5c3a310ddaa2c1280c3ec3ff20214e3854b46e6c2675ba1c82ea455\": container with ID starting with 216d66fdf5c3a310ddaa2c1280c3ec3ff20214e3854b46e6c2675ba1c82ea455 not found: ID does not exist" Dec 10 20:27:45 crc kubenswrapper[4828]: I1210 20:27:45.805779 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7848156-6783-4de5-a9cb-07f89c916747" path="/var/lib/kubelet/pods/c7848156-6783-4de5-a9cb-07f89c916747/volumes" Dec 10 20:28:09 crc kubenswrapper[4828]: I1210 20:28:09.785261 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2vlz9"] Dec 10 20:28:09 crc kubenswrapper[4828]: E1210 20:28:09.787442 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7848156-6783-4de5-a9cb-07f89c916747" containerName="registry-server" Dec 10 20:28:09 crc kubenswrapper[4828]: I1210 20:28:09.787525 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7848156-6783-4de5-a9cb-07f89c916747" containerName="registry-server" Dec 10 20:28:09 crc kubenswrapper[4828]: E1210 20:28:09.787591 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7848156-6783-4de5-a9cb-07f89c916747" containerName="extract-utilities" Dec 10 20:28:09 crc kubenswrapper[4828]: I1210 20:28:09.787647 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7848156-6783-4de5-a9cb-07f89c916747" containerName="extract-utilities" Dec 10 20:28:09 crc kubenswrapper[4828]: E1210 20:28:09.787711 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7848156-6783-4de5-a9cb-07f89c916747" containerName="extract-content" Dec 10 20:28:09 crc kubenswrapper[4828]: I1210 20:28:09.787765 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7848156-6783-4de5-a9cb-07f89c916747" containerName="extract-content" Dec 10 20:28:09 crc kubenswrapper[4828]: I1210 20:28:09.788095 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7848156-6783-4de5-a9cb-07f89c916747" containerName="registry-server" Dec 10 20:28:09 crc kubenswrapper[4828]: I1210 20:28:09.791106 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2vlz9" Dec 10 20:28:09 crc kubenswrapper[4828]: I1210 20:28:09.811481 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2vlz9"] Dec 10 20:28:09 crc kubenswrapper[4828]: I1210 20:28:09.938108 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xn5m\" (UniqueName: \"kubernetes.io/projected/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-kube-api-access-7xn5m\") pod \"redhat-marketplace-2vlz9\" (UID: \"9f66cafa-27ce-4b3e-a789-d7e82242d0fe\") " pod="openshift-marketplace/redhat-marketplace-2vlz9" Dec 10 20:28:09 crc kubenswrapper[4828]: I1210 20:28:09.938482 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-catalog-content\") pod \"redhat-marketplace-2vlz9\" (UID: \"9f66cafa-27ce-4b3e-a789-d7e82242d0fe\") " pod="openshift-marketplace/redhat-marketplace-2vlz9" Dec 10 20:28:09 crc kubenswrapper[4828]: I1210 20:28:09.939006 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-utilities\") pod \"redhat-marketplace-2vlz9\" (UID: \"9f66cafa-27ce-4b3e-a789-d7e82242d0fe\") " pod="openshift-marketplace/redhat-marketplace-2vlz9" Dec 10 20:28:10 crc kubenswrapper[4828]: I1210 20:28:10.041603 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-utilities\") pod \"redhat-marketplace-2vlz9\" (UID: \"9f66cafa-27ce-4b3e-a789-d7e82242d0fe\") " pod="openshift-marketplace/redhat-marketplace-2vlz9" Dec 10 20:28:10 crc kubenswrapper[4828]: I1210 20:28:10.041853 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xn5m\" (UniqueName: \"kubernetes.io/projected/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-kube-api-access-7xn5m\") pod \"redhat-marketplace-2vlz9\" (UID: \"9f66cafa-27ce-4b3e-a789-d7e82242d0fe\") " pod="openshift-marketplace/redhat-marketplace-2vlz9" Dec 10 20:28:10 crc kubenswrapper[4828]: I1210 20:28:10.042010 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-catalog-content\") pod \"redhat-marketplace-2vlz9\" (UID: \"9f66cafa-27ce-4b3e-a789-d7e82242d0fe\") " pod="openshift-marketplace/redhat-marketplace-2vlz9" Dec 10 20:28:10 crc kubenswrapper[4828]: I1210 20:28:10.042221 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-utilities\") pod \"redhat-marketplace-2vlz9\" (UID: \"9f66cafa-27ce-4b3e-a789-d7e82242d0fe\") " pod="openshift-marketplace/redhat-marketplace-2vlz9" Dec 10 20:28:10 crc kubenswrapper[4828]: I1210 20:28:10.042619 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-catalog-content\") pod \"redhat-marketplace-2vlz9\" (UID: \"9f66cafa-27ce-4b3e-a789-d7e82242d0fe\") " pod="openshift-marketplace/redhat-marketplace-2vlz9" Dec 10 20:28:10 crc kubenswrapper[4828]: I1210 20:28:10.070076 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xn5m\" (UniqueName: \"kubernetes.io/projected/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-kube-api-access-7xn5m\") pod \"redhat-marketplace-2vlz9\" (UID: \"9f66cafa-27ce-4b3e-a789-d7e82242d0fe\") " pod="openshift-marketplace/redhat-marketplace-2vlz9" Dec 10 20:28:10 crc kubenswrapper[4828]: I1210 20:28:10.130509 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2vlz9" Dec 10 20:28:10 crc kubenswrapper[4828]: I1210 20:28:10.611429 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2vlz9"] Dec 10 20:28:11 crc kubenswrapper[4828]: I1210 20:28:11.103102 4828 generic.go:334] "Generic (PLEG): container finished" podID="9f66cafa-27ce-4b3e-a789-d7e82242d0fe" containerID="ce7ea04938829b931b9a8fed4b520aa2ffbfc5f2da2debbdfb87d2d9a7f8bfb1" exitCode=0 Dec 10 20:28:11 crc kubenswrapper[4828]: I1210 20:28:11.103155 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2vlz9" event={"ID":"9f66cafa-27ce-4b3e-a789-d7e82242d0fe","Type":"ContainerDied","Data":"ce7ea04938829b931b9a8fed4b520aa2ffbfc5f2da2debbdfb87d2d9a7f8bfb1"} Dec 10 20:28:11 crc kubenswrapper[4828]: I1210 20:28:11.103434 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2vlz9" event={"ID":"9f66cafa-27ce-4b3e-a789-d7e82242d0fe","Type":"ContainerStarted","Data":"67a8afadee93f84c2083d9c2bc17f938cbd81a0f88c23dc6ea4ff60e5ec3e25d"} Dec 10 20:28:13 crc kubenswrapper[4828]: I1210 20:28:13.124580 4828 generic.go:334] "Generic (PLEG): container finished" podID="9f66cafa-27ce-4b3e-a789-d7e82242d0fe" containerID="e16f458f22dcf1723420818e2157f1b5a7acf035618c6bb9c145b393ee45582f" exitCode=0 Dec 10 20:28:13 crc kubenswrapper[4828]: I1210 20:28:13.124621 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2vlz9" event={"ID":"9f66cafa-27ce-4b3e-a789-d7e82242d0fe","Type":"ContainerDied","Data":"e16f458f22dcf1723420818e2157f1b5a7acf035618c6bb9c145b393ee45582f"} Dec 10 20:28:14 crc kubenswrapper[4828]: I1210 20:28:14.162858 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2vlz9" event={"ID":"9f66cafa-27ce-4b3e-a789-d7e82242d0fe","Type":"ContainerStarted","Data":"3428e00ff6d5cf5cee2ddb525cfbff5aefdb42e4f152fc37876a39fe793738e3"} Dec 10 20:28:14 crc kubenswrapper[4828]: I1210 20:28:14.190732 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2vlz9" podStartSLOduration=2.680908701 podStartE2EDuration="5.190713202s" podCreationTimestamp="2025-12-10 20:28:09 +0000 UTC" firstStartedPulling="2025-12-10 20:28:11.106232505 +0000 UTC m=+5571.616843510" lastFinishedPulling="2025-12-10 20:28:13.616037006 +0000 UTC m=+5574.126648011" observedRunningTime="2025-12-10 20:28:14.184495136 +0000 UTC m=+5574.695106161" watchObservedRunningTime="2025-12-10 20:28:14.190713202 +0000 UTC m=+5574.701324207" Dec 10 20:28:20 crc kubenswrapper[4828]: I1210 20:28:20.131104 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2vlz9" Dec 10 20:28:20 crc kubenswrapper[4828]: I1210 20:28:20.132477 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2vlz9" Dec 10 20:28:20 crc kubenswrapper[4828]: I1210 20:28:20.932530 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2vlz9" Dec 10 20:28:21 crc kubenswrapper[4828]: I1210 20:28:21.010983 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2vlz9" Dec 10 20:28:21 crc kubenswrapper[4828]: I1210 20:28:21.182449 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2vlz9"] Dec 10 20:28:22 crc kubenswrapper[4828]: I1210 20:28:22.261520 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2vlz9" podUID="9f66cafa-27ce-4b3e-a789-d7e82242d0fe" containerName="registry-server" containerID="cri-o://3428e00ff6d5cf5cee2ddb525cfbff5aefdb42e4f152fc37876a39fe793738e3" gracePeriod=2 Dec 10 20:28:22 crc kubenswrapper[4828]: I1210 20:28:22.839970 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2vlz9" Dec 10 20:28:22 crc kubenswrapper[4828]: I1210 20:28:22.925785 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-utilities\") pod \"9f66cafa-27ce-4b3e-a789-d7e82242d0fe\" (UID: \"9f66cafa-27ce-4b3e-a789-d7e82242d0fe\") " Dec 10 20:28:22 crc kubenswrapper[4828]: I1210 20:28:22.925899 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-catalog-content\") pod \"9f66cafa-27ce-4b3e-a789-d7e82242d0fe\" (UID: \"9f66cafa-27ce-4b3e-a789-d7e82242d0fe\") " Dec 10 20:28:22 crc kubenswrapper[4828]: I1210 20:28:22.926060 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xn5m\" (UniqueName: \"kubernetes.io/projected/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-kube-api-access-7xn5m\") pod \"9f66cafa-27ce-4b3e-a789-d7e82242d0fe\" (UID: \"9f66cafa-27ce-4b3e-a789-d7e82242d0fe\") " Dec 10 20:28:22 crc kubenswrapper[4828]: I1210 20:28:22.926873 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-utilities" (OuterVolumeSpecName: "utilities") pod "9f66cafa-27ce-4b3e-a789-d7e82242d0fe" (UID: "9f66cafa-27ce-4b3e-a789-d7e82242d0fe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:28:22 crc kubenswrapper[4828]: I1210 20:28:22.932671 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-kube-api-access-7xn5m" (OuterVolumeSpecName: "kube-api-access-7xn5m") pod "9f66cafa-27ce-4b3e-a789-d7e82242d0fe" (UID: "9f66cafa-27ce-4b3e-a789-d7e82242d0fe"). InnerVolumeSpecName "kube-api-access-7xn5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:28:22 crc kubenswrapper[4828]: I1210 20:28:22.946337 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9f66cafa-27ce-4b3e-a789-d7e82242d0fe" (UID: "9f66cafa-27ce-4b3e-a789-d7e82242d0fe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.028658 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.028703 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xn5m\" (UniqueName: \"kubernetes.io/projected/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-kube-api-access-7xn5m\") on node \"crc\" DevicePath \"\"" Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.028715 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f66cafa-27ce-4b3e-a789-d7e82242d0fe-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.270545 4828 generic.go:334] "Generic (PLEG): container finished" podID="9f66cafa-27ce-4b3e-a789-d7e82242d0fe" containerID="3428e00ff6d5cf5cee2ddb525cfbff5aefdb42e4f152fc37876a39fe793738e3" exitCode=0 Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.270602 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2vlz9" Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.270627 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2vlz9" event={"ID":"9f66cafa-27ce-4b3e-a789-d7e82242d0fe","Type":"ContainerDied","Data":"3428e00ff6d5cf5cee2ddb525cfbff5aefdb42e4f152fc37876a39fe793738e3"} Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.270973 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2vlz9" event={"ID":"9f66cafa-27ce-4b3e-a789-d7e82242d0fe","Type":"ContainerDied","Data":"67a8afadee93f84c2083d9c2bc17f938cbd81a0f88c23dc6ea4ff60e5ec3e25d"} Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.270991 4828 scope.go:117] "RemoveContainer" containerID="3428e00ff6d5cf5cee2ddb525cfbff5aefdb42e4f152fc37876a39fe793738e3" Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.298980 4828 scope.go:117] "RemoveContainer" containerID="e16f458f22dcf1723420818e2157f1b5a7acf035618c6bb9c145b393ee45582f" Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.308287 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2vlz9"] Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.327825 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2vlz9"] Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.330073 4828 scope.go:117] "RemoveContainer" containerID="ce7ea04938829b931b9a8fed4b520aa2ffbfc5f2da2debbdfb87d2d9a7f8bfb1" Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.374763 4828 scope.go:117] "RemoveContainer" containerID="3428e00ff6d5cf5cee2ddb525cfbff5aefdb42e4f152fc37876a39fe793738e3" Dec 10 20:28:23 crc kubenswrapper[4828]: E1210 20:28:23.375174 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3428e00ff6d5cf5cee2ddb525cfbff5aefdb42e4f152fc37876a39fe793738e3\": container with ID starting with 3428e00ff6d5cf5cee2ddb525cfbff5aefdb42e4f152fc37876a39fe793738e3 not found: ID does not exist" containerID="3428e00ff6d5cf5cee2ddb525cfbff5aefdb42e4f152fc37876a39fe793738e3" Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.375210 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3428e00ff6d5cf5cee2ddb525cfbff5aefdb42e4f152fc37876a39fe793738e3"} err="failed to get container status \"3428e00ff6d5cf5cee2ddb525cfbff5aefdb42e4f152fc37876a39fe793738e3\": rpc error: code = NotFound desc = could not find container \"3428e00ff6d5cf5cee2ddb525cfbff5aefdb42e4f152fc37876a39fe793738e3\": container with ID starting with 3428e00ff6d5cf5cee2ddb525cfbff5aefdb42e4f152fc37876a39fe793738e3 not found: ID does not exist" Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.375231 4828 scope.go:117] "RemoveContainer" containerID="e16f458f22dcf1723420818e2157f1b5a7acf035618c6bb9c145b393ee45582f" Dec 10 20:28:23 crc kubenswrapper[4828]: E1210 20:28:23.375451 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e16f458f22dcf1723420818e2157f1b5a7acf035618c6bb9c145b393ee45582f\": container with ID starting with e16f458f22dcf1723420818e2157f1b5a7acf035618c6bb9c145b393ee45582f not found: ID does not exist" containerID="e16f458f22dcf1723420818e2157f1b5a7acf035618c6bb9c145b393ee45582f" Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.375497 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e16f458f22dcf1723420818e2157f1b5a7acf035618c6bb9c145b393ee45582f"} err="failed to get container status \"e16f458f22dcf1723420818e2157f1b5a7acf035618c6bb9c145b393ee45582f\": rpc error: code = NotFound desc = could not find container \"e16f458f22dcf1723420818e2157f1b5a7acf035618c6bb9c145b393ee45582f\": container with ID starting with e16f458f22dcf1723420818e2157f1b5a7acf035618c6bb9c145b393ee45582f not found: ID does not exist" Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.375514 4828 scope.go:117] "RemoveContainer" containerID="ce7ea04938829b931b9a8fed4b520aa2ffbfc5f2da2debbdfb87d2d9a7f8bfb1" Dec 10 20:28:23 crc kubenswrapper[4828]: E1210 20:28:23.375788 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce7ea04938829b931b9a8fed4b520aa2ffbfc5f2da2debbdfb87d2d9a7f8bfb1\": container with ID starting with ce7ea04938829b931b9a8fed4b520aa2ffbfc5f2da2debbdfb87d2d9a7f8bfb1 not found: ID does not exist" containerID="ce7ea04938829b931b9a8fed4b520aa2ffbfc5f2da2debbdfb87d2d9a7f8bfb1" Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.375837 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce7ea04938829b931b9a8fed4b520aa2ffbfc5f2da2debbdfb87d2d9a7f8bfb1"} err="failed to get container status \"ce7ea04938829b931b9a8fed4b520aa2ffbfc5f2da2debbdfb87d2d9a7f8bfb1\": rpc error: code = NotFound desc = could not find container \"ce7ea04938829b931b9a8fed4b520aa2ffbfc5f2da2debbdfb87d2d9a7f8bfb1\": container with ID starting with ce7ea04938829b931b9a8fed4b520aa2ffbfc5f2da2debbdfb87d2d9a7f8bfb1 not found: ID does not exist" Dec 10 20:28:23 crc kubenswrapper[4828]: I1210 20:28:23.805971 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f66cafa-27ce-4b3e-a789-d7e82242d0fe" path="/var/lib/kubelet/pods/9f66cafa-27ce-4b3e-a789-d7e82242d0fe/volumes" Dec 10 20:28:51 crc kubenswrapper[4828]: I1210 20:28:51.230950 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:28:51 crc kubenswrapper[4828]: I1210 20:28:51.231745 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:29:21 crc kubenswrapper[4828]: I1210 20:29:21.230867 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:29:21 crc kubenswrapper[4828]: I1210 20:29:21.231313 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:29:35 crc kubenswrapper[4828]: I1210 20:29:35.109599 4828 generic.go:334] "Generic (PLEG): container finished" podID="169c0824-0503-4c49-9980-7d045db64b8b" containerID="b0ded671bccc9c9ec6244ca39ec5ad6af9608530e8dce8290ac1dc4cf7764f65" exitCode=0 Dec 10 20:29:35 crc kubenswrapper[4828]: I1210 20:29:35.109694 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"169c0824-0503-4c49-9980-7d045db64b8b","Type":"ContainerDied","Data":"b0ded671bccc9c9ec6244ca39ec5ad6af9608530e8dce8290ac1dc4cf7764f65"} Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.530881 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.601914 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/169c0824-0503-4c49-9980-7d045db64b8b-config-data\") pod \"169c0824-0503-4c49-9980-7d045db64b8b\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.601973 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/169c0824-0503-4c49-9980-7d045db64b8b-test-operator-ephemeral-temporary\") pod \"169c0824-0503-4c49-9980-7d045db64b8b\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.602035 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/169c0824-0503-4c49-9980-7d045db64b8b-test-operator-ephemeral-workdir\") pod \"169c0824-0503-4c49-9980-7d045db64b8b\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.602104 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-ssh-key\") pod \"169c0824-0503-4c49-9980-7d045db64b8b\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.602323 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/169c0824-0503-4c49-9980-7d045db64b8b-openstack-config\") pod \"169c0824-0503-4c49-9980-7d045db64b8b\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.602420 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-openstack-config-secret\") pod \"169c0824-0503-4c49-9980-7d045db64b8b\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.602458 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-ca-certs\") pod \"169c0824-0503-4c49-9980-7d045db64b8b\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.602482 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7hvt\" (UniqueName: \"kubernetes.io/projected/169c0824-0503-4c49-9980-7d045db64b8b-kube-api-access-g7hvt\") pod \"169c0824-0503-4c49-9980-7d045db64b8b\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.602504 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"169c0824-0503-4c49-9980-7d045db64b8b\" (UID: \"169c0824-0503-4c49-9980-7d045db64b8b\") " Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.602845 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/169c0824-0503-4c49-9980-7d045db64b8b-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "169c0824-0503-4c49-9980-7d045db64b8b" (UID: "169c0824-0503-4c49-9980-7d045db64b8b"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.603222 4828 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/169c0824-0503-4c49-9980-7d045db64b8b-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.603861 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/169c0824-0503-4c49-9980-7d045db64b8b-config-data" (OuterVolumeSpecName: "config-data") pod "169c0824-0503-4c49-9980-7d045db64b8b" (UID: "169c0824-0503-4c49-9980-7d045db64b8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.609599 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "test-operator-logs") pod "169c0824-0503-4c49-9980-7d045db64b8b" (UID: "169c0824-0503-4c49-9980-7d045db64b8b"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.613045 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/169c0824-0503-4c49-9980-7d045db64b8b-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "169c0824-0503-4c49-9980-7d045db64b8b" (UID: "169c0824-0503-4c49-9980-7d045db64b8b"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.613599 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/169c0824-0503-4c49-9980-7d045db64b8b-kube-api-access-g7hvt" (OuterVolumeSpecName: "kube-api-access-g7hvt") pod "169c0824-0503-4c49-9980-7d045db64b8b" (UID: "169c0824-0503-4c49-9980-7d045db64b8b"). InnerVolumeSpecName "kube-api-access-g7hvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.638893 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "169c0824-0503-4c49-9980-7d045db64b8b" (UID: "169c0824-0503-4c49-9980-7d045db64b8b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.639713 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "169c0824-0503-4c49-9980-7d045db64b8b" (UID: "169c0824-0503-4c49-9980-7d045db64b8b"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.641617 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "169c0824-0503-4c49-9980-7d045db64b8b" (UID: "169c0824-0503-4c49-9980-7d045db64b8b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.664873 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/169c0824-0503-4c49-9980-7d045db64b8b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "169c0824-0503-4c49-9980-7d045db64b8b" (UID: "169c0824-0503-4c49-9980-7d045db64b8b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.705490 4828 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/169c0824-0503-4c49-9980-7d045db64b8b-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.705792 4828 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.705895 4828 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.705981 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7hvt\" (UniqueName: \"kubernetes.io/projected/169c0824-0503-4c49-9980-7d045db64b8b-kube-api-access-g7hvt\") on node \"crc\" DevicePath \"\"" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.706240 4828 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.706317 4828 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/169c0824-0503-4c49-9980-7d045db64b8b-config-data\") on node \"crc\" DevicePath \"\"" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.706378 4828 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/169c0824-0503-4c49-9980-7d045db64b8b-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.706433 4828 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/169c0824-0503-4c49-9980-7d045db64b8b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.732904 4828 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 10 20:29:36 crc kubenswrapper[4828]: I1210 20:29:36.809182 4828 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 10 20:29:37 crc kubenswrapper[4828]: I1210 20:29:37.134358 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"169c0824-0503-4c49-9980-7d045db64b8b","Type":"ContainerDied","Data":"741f6fa6ea23b23dccef289a99c626be723d7212950ad01faf1eb6a631d116ad"} Dec 10 20:29:37 crc kubenswrapper[4828]: I1210 20:29:37.134399 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="741f6fa6ea23b23dccef289a99c626be723d7212950ad01faf1eb6a631d116ad" Dec 10 20:29:37 crc kubenswrapper[4828]: I1210 20:29:37.134409 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 10 20:29:43 crc kubenswrapper[4828]: I1210 20:29:43.918771 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 10 20:29:43 crc kubenswrapper[4828]: E1210 20:29:43.919786 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f66cafa-27ce-4b3e-a789-d7e82242d0fe" containerName="extract-content" Dec 10 20:29:43 crc kubenswrapper[4828]: I1210 20:29:43.919827 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f66cafa-27ce-4b3e-a789-d7e82242d0fe" containerName="extract-content" Dec 10 20:29:43 crc kubenswrapper[4828]: E1210 20:29:43.919870 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="169c0824-0503-4c49-9980-7d045db64b8b" containerName="tempest-tests-tempest-tests-runner" Dec 10 20:29:43 crc kubenswrapper[4828]: I1210 20:29:43.919876 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="169c0824-0503-4c49-9980-7d045db64b8b" containerName="tempest-tests-tempest-tests-runner" Dec 10 20:29:43 crc kubenswrapper[4828]: E1210 20:29:43.919889 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f66cafa-27ce-4b3e-a789-d7e82242d0fe" containerName="extract-utilities" Dec 10 20:29:43 crc kubenswrapper[4828]: I1210 20:29:43.919895 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f66cafa-27ce-4b3e-a789-d7e82242d0fe" containerName="extract-utilities" Dec 10 20:29:43 crc kubenswrapper[4828]: E1210 20:29:43.919911 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f66cafa-27ce-4b3e-a789-d7e82242d0fe" containerName="registry-server" Dec 10 20:29:43 crc kubenswrapper[4828]: I1210 20:29:43.919917 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f66cafa-27ce-4b3e-a789-d7e82242d0fe" containerName="registry-server" Dec 10 20:29:43 crc kubenswrapper[4828]: I1210 20:29:43.920106 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f66cafa-27ce-4b3e-a789-d7e82242d0fe" containerName="registry-server" Dec 10 20:29:43 crc kubenswrapper[4828]: I1210 20:29:43.920133 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="169c0824-0503-4c49-9980-7d045db64b8b" containerName="tempest-tests-tempest-tests-runner" Dec 10 20:29:43 crc kubenswrapper[4828]: I1210 20:29:43.920977 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 20:29:43 crc kubenswrapper[4828]: I1210 20:29:43.926536 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-zghwg" Dec 10 20:29:43 crc kubenswrapper[4828]: I1210 20:29:43.934429 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 10 20:29:44 crc kubenswrapper[4828]: I1210 20:29:44.097206 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"11bbb31c-6748-4445-aa7d-bebfb5578925\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 20:29:44 crc kubenswrapper[4828]: I1210 20:29:44.097608 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmgqd\" (UniqueName: \"kubernetes.io/projected/11bbb31c-6748-4445-aa7d-bebfb5578925-kube-api-access-cmgqd\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"11bbb31c-6748-4445-aa7d-bebfb5578925\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 20:29:44 crc kubenswrapper[4828]: I1210 20:29:44.200128 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmgqd\" (UniqueName: \"kubernetes.io/projected/11bbb31c-6748-4445-aa7d-bebfb5578925-kube-api-access-cmgqd\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"11bbb31c-6748-4445-aa7d-bebfb5578925\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 20:29:44 crc kubenswrapper[4828]: I1210 20:29:44.200300 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"11bbb31c-6748-4445-aa7d-bebfb5578925\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 20:29:44 crc kubenswrapper[4828]: I1210 20:29:44.201395 4828 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"11bbb31c-6748-4445-aa7d-bebfb5578925\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 20:29:44 crc kubenswrapper[4828]: I1210 20:29:44.224442 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmgqd\" (UniqueName: \"kubernetes.io/projected/11bbb31c-6748-4445-aa7d-bebfb5578925-kube-api-access-cmgqd\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"11bbb31c-6748-4445-aa7d-bebfb5578925\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 20:29:44 crc kubenswrapper[4828]: I1210 20:29:44.235551 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"11bbb31c-6748-4445-aa7d-bebfb5578925\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 20:29:44 crc kubenswrapper[4828]: I1210 20:29:44.253926 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 10 20:29:44 crc kubenswrapper[4828]: I1210 20:29:44.723196 4828 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 20:29:44 crc kubenswrapper[4828]: I1210 20:29:44.723426 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 10 20:29:45 crc kubenswrapper[4828]: I1210 20:29:45.226846 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"11bbb31c-6748-4445-aa7d-bebfb5578925","Type":"ContainerStarted","Data":"f5b5e7973d194c6f8ed0f15170bf64ec23e3137c0c6e0ed4db9e49f2e16eb165"} Dec 10 20:29:47 crc kubenswrapper[4828]: I1210 20:29:47.248230 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"11bbb31c-6748-4445-aa7d-bebfb5578925","Type":"ContainerStarted","Data":"d18a6e10ae8368e47bc30c0079f27a62017f2dd7fb72fe5a4b07b170b451b1eb"} Dec 10 20:29:47 crc kubenswrapper[4828]: I1210 20:29:47.264257 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=3.214533291 podStartE2EDuration="4.264234057s" podCreationTimestamp="2025-12-10 20:29:43 +0000 UTC" firstStartedPulling="2025-12-10 20:29:44.723005785 +0000 UTC m=+5665.233616790" lastFinishedPulling="2025-12-10 20:29:45.772706541 +0000 UTC m=+5666.283317556" observedRunningTime="2025-12-10 20:29:47.260791215 +0000 UTC m=+5667.771402220" watchObservedRunningTime="2025-12-10 20:29:47.264234057 +0000 UTC m=+5667.774845062" Dec 10 20:29:51 crc kubenswrapper[4828]: I1210 20:29:51.231081 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:29:51 crc kubenswrapper[4828]: I1210 20:29:51.231874 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:29:51 crc kubenswrapper[4828]: I1210 20:29:51.231936 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 20:29:51 crc kubenswrapper[4828]: I1210 20:29:51.232981 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 20:29:51 crc kubenswrapper[4828]: I1210 20:29:51.233066 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" gracePeriod=600 Dec 10 20:29:51 crc kubenswrapper[4828]: E1210 20:29:51.351734 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:29:51 crc kubenswrapper[4828]: I1210 20:29:51.392890 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" exitCode=0 Dec 10 20:29:51 crc kubenswrapper[4828]: I1210 20:29:51.392963 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2"} Dec 10 20:29:51 crc kubenswrapper[4828]: I1210 20:29:51.393031 4828 scope.go:117] "RemoveContainer" containerID="3886f27f4fecd3d3e16c0bed1ffc46840f19c5699ccc27978965f246e4eaad13" Dec 10 20:29:51 crc kubenswrapper[4828]: I1210 20:29:51.394989 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:29:51 crc kubenswrapper[4828]: E1210 20:29:51.395910 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:30:00 crc kubenswrapper[4828]: I1210 20:30:00.173370 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb"] Dec 10 20:30:00 crc kubenswrapper[4828]: I1210 20:30:00.176134 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb" Dec 10 20:30:00 crc kubenswrapper[4828]: I1210 20:30:00.180072 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 10 20:30:00 crc kubenswrapper[4828]: I1210 20:30:00.180447 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 10 20:30:00 crc kubenswrapper[4828]: I1210 20:30:00.191426 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb"] Dec 10 20:30:00 crc kubenswrapper[4828]: I1210 20:30:00.292183 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db4c3816-9d19-4286-9694-4de74f179779-config-volume\") pod \"collect-profiles-29423310-548jb\" (UID: \"db4c3816-9d19-4286-9694-4de74f179779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb" Dec 10 20:30:00 crc kubenswrapper[4828]: I1210 20:30:00.292565 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db4c3816-9d19-4286-9694-4de74f179779-secret-volume\") pod \"collect-profiles-29423310-548jb\" (UID: \"db4c3816-9d19-4286-9694-4de74f179779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb" Dec 10 20:30:00 crc kubenswrapper[4828]: I1210 20:30:00.292622 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx2x7\" (UniqueName: \"kubernetes.io/projected/db4c3816-9d19-4286-9694-4de74f179779-kube-api-access-dx2x7\") pod \"collect-profiles-29423310-548jb\" (UID: \"db4c3816-9d19-4286-9694-4de74f179779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb" Dec 10 20:30:00 crc kubenswrapper[4828]: I1210 20:30:00.395075 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx2x7\" (UniqueName: \"kubernetes.io/projected/db4c3816-9d19-4286-9694-4de74f179779-kube-api-access-dx2x7\") pod \"collect-profiles-29423310-548jb\" (UID: \"db4c3816-9d19-4286-9694-4de74f179779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb" Dec 10 20:30:00 crc kubenswrapper[4828]: I1210 20:30:00.395207 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db4c3816-9d19-4286-9694-4de74f179779-config-volume\") pod \"collect-profiles-29423310-548jb\" (UID: \"db4c3816-9d19-4286-9694-4de74f179779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb" Dec 10 20:30:00 crc kubenswrapper[4828]: I1210 20:30:00.395348 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db4c3816-9d19-4286-9694-4de74f179779-secret-volume\") pod \"collect-profiles-29423310-548jb\" (UID: \"db4c3816-9d19-4286-9694-4de74f179779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb" Dec 10 20:30:00 crc kubenswrapper[4828]: I1210 20:30:00.396011 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db4c3816-9d19-4286-9694-4de74f179779-config-volume\") pod \"collect-profiles-29423310-548jb\" (UID: \"db4c3816-9d19-4286-9694-4de74f179779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb" Dec 10 20:30:00 crc kubenswrapper[4828]: I1210 20:30:00.409817 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db4c3816-9d19-4286-9694-4de74f179779-secret-volume\") pod \"collect-profiles-29423310-548jb\" (UID: \"db4c3816-9d19-4286-9694-4de74f179779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb" Dec 10 20:30:00 crc kubenswrapper[4828]: I1210 20:30:00.417012 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx2x7\" (UniqueName: \"kubernetes.io/projected/db4c3816-9d19-4286-9694-4de74f179779-kube-api-access-dx2x7\") pod \"collect-profiles-29423310-548jb\" (UID: \"db4c3816-9d19-4286-9694-4de74f179779\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb" Dec 10 20:30:00 crc kubenswrapper[4828]: I1210 20:30:00.509488 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb" Dec 10 20:30:00 crc kubenswrapper[4828]: I1210 20:30:00.987360 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb"] Dec 10 20:30:00 crc kubenswrapper[4828]: W1210 20:30:00.988372 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb4c3816_9d19_4286_9694_4de74f179779.slice/crio-5f900c500cc94675c3f4d5fe7ca677a8cdaa31346eb7e6607afc976afad158d2 WatchSource:0}: Error finding container 5f900c500cc94675c3f4d5fe7ca677a8cdaa31346eb7e6607afc976afad158d2: Status 404 returned error can't find the container with id 5f900c500cc94675c3f4d5fe7ca677a8cdaa31346eb7e6607afc976afad158d2 Dec 10 20:30:01 crc kubenswrapper[4828]: I1210 20:30:01.518162 4828 generic.go:334] "Generic (PLEG): container finished" podID="db4c3816-9d19-4286-9694-4de74f179779" containerID="43d2f9387e0a03d29db4cff1595c515530d7db64e5bcbc7aaf7746517321b665" exitCode=0 Dec 10 20:30:01 crc kubenswrapper[4828]: I1210 20:30:01.518364 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb" event={"ID":"db4c3816-9d19-4286-9694-4de74f179779","Type":"ContainerDied","Data":"43d2f9387e0a03d29db4cff1595c515530d7db64e5bcbc7aaf7746517321b665"} Dec 10 20:30:01 crc kubenswrapper[4828]: I1210 20:30:01.518473 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb" event={"ID":"db4c3816-9d19-4286-9694-4de74f179779","Type":"ContainerStarted","Data":"5f900c500cc94675c3f4d5fe7ca677a8cdaa31346eb7e6607afc976afad158d2"} Dec 10 20:30:02 crc kubenswrapper[4828]: I1210 20:30:02.914198 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb" Dec 10 20:30:02 crc kubenswrapper[4828]: I1210 20:30:02.974899 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db4c3816-9d19-4286-9694-4de74f179779-secret-volume\") pod \"db4c3816-9d19-4286-9694-4de74f179779\" (UID: \"db4c3816-9d19-4286-9694-4de74f179779\") " Dec 10 20:30:02 crc kubenswrapper[4828]: I1210 20:30:02.975051 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dx2x7\" (UniqueName: \"kubernetes.io/projected/db4c3816-9d19-4286-9694-4de74f179779-kube-api-access-dx2x7\") pod \"db4c3816-9d19-4286-9694-4de74f179779\" (UID: \"db4c3816-9d19-4286-9694-4de74f179779\") " Dec 10 20:30:02 crc kubenswrapper[4828]: I1210 20:30:02.975150 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db4c3816-9d19-4286-9694-4de74f179779-config-volume\") pod \"db4c3816-9d19-4286-9694-4de74f179779\" (UID: \"db4c3816-9d19-4286-9694-4de74f179779\") " Dec 10 20:30:02 crc kubenswrapper[4828]: I1210 20:30:02.977935 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db4c3816-9d19-4286-9694-4de74f179779-config-volume" (OuterVolumeSpecName: "config-volume") pod "db4c3816-9d19-4286-9694-4de74f179779" (UID: "db4c3816-9d19-4286-9694-4de74f179779"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 10 20:30:02 crc kubenswrapper[4828]: I1210 20:30:02.985256 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db4c3816-9d19-4286-9694-4de74f179779-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "db4c3816-9d19-4286-9694-4de74f179779" (UID: "db4c3816-9d19-4286-9694-4de74f179779"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 10 20:30:02 crc kubenswrapper[4828]: I1210 20:30:02.995208 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db4c3816-9d19-4286-9694-4de74f179779-kube-api-access-dx2x7" (OuterVolumeSpecName: "kube-api-access-dx2x7") pod "db4c3816-9d19-4286-9694-4de74f179779" (UID: "db4c3816-9d19-4286-9694-4de74f179779"). InnerVolumeSpecName "kube-api-access-dx2x7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:30:03 crc kubenswrapper[4828]: I1210 20:30:03.080660 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dx2x7\" (UniqueName: \"kubernetes.io/projected/db4c3816-9d19-4286-9694-4de74f179779-kube-api-access-dx2x7\") on node \"crc\" DevicePath \"\"" Dec 10 20:30:03 crc kubenswrapper[4828]: I1210 20:30:03.080694 4828 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db4c3816-9d19-4286-9694-4de74f179779-config-volume\") on node \"crc\" DevicePath \"\"" Dec 10 20:30:03 crc kubenswrapper[4828]: I1210 20:30:03.080703 4828 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db4c3816-9d19-4286-9694-4de74f179779-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 10 20:30:03 crc kubenswrapper[4828]: I1210 20:30:03.544082 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb" event={"ID":"db4c3816-9d19-4286-9694-4de74f179779","Type":"ContainerDied","Data":"5f900c500cc94675c3f4d5fe7ca677a8cdaa31346eb7e6607afc976afad158d2"} Dec 10 20:30:03 crc kubenswrapper[4828]: I1210 20:30:03.544123 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f900c500cc94675c3f4d5fe7ca677a8cdaa31346eb7e6607afc976afad158d2" Dec 10 20:30:03 crc kubenswrapper[4828]: I1210 20:30:03.544146 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29423310-548jb" Dec 10 20:30:03 crc kubenswrapper[4828]: I1210 20:30:03.983684 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr"] Dec 10 20:30:03 crc kubenswrapper[4828]: I1210 20:30:03.995318 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29423265-s7hpr"] Dec 10 20:30:05 crc kubenswrapper[4828]: I1210 20:30:05.789434 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:30:05 crc kubenswrapper[4828]: E1210 20:30:05.790150 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:30:05 crc kubenswrapper[4828]: I1210 20:30:05.807577 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f445c4c6-49ea-483f-b20c-22fb12e316ad" path="/var/lib/kubelet/pods/f445c4c6-49ea-483f-b20c-22fb12e316ad/volumes" Dec 10 20:30:18 crc kubenswrapper[4828]: I1210 20:30:18.788564 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:30:18 crc kubenswrapper[4828]: E1210 20:30:18.789588 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:30:31 crc kubenswrapper[4828]: I1210 20:30:31.789778 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:30:31 crc kubenswrapper[4828]: E1210 20:30:31.790647 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:30:32 crc kubenswrapper[4828]: I1210 20:30:32.135783 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qpwwn/must-gather-n5s68"] Dec 10 20:30:32 crc kubenswrapper[4828]: E1210 20:30:32.136659 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db4c3816-9d19-4286-9694-4de74f179779" containerName="collect-profiles" Dec 10 20:30:32 crc kubenswrapper[4828]: I1210 20:30:32.136677 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="db4c3816-9d19-4286-9694-4de74f179779" containerName="collect-profiles" Dec 10 20:30:32 crc kubenswrapper[4828]: I1210 20:30:32.136939 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="db4c3816-9d19-4286-9694-4de74f179779" containerName="collect-profiles" Dec 10 20:30:32 crc kubenswrapper[4828]: I1210 20:30:32.138392 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qpwwn/must-gather-n5s68" Dec 10 20:30:32 crc kubenswrapper[4828]: I1210 20:30:32.140472 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qpwwn"/"openshift-service-ca.crt" Dec 10 20:30:32 crc kubenswrapper[4828]: I1210 20:30:32.140587 4828 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qpwwn"/"default-dockercfg-t62bd" Dec 10 20:30:32 crc kubenswrapper[4828]: I1210 20:30:32.140728 4828 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qpwwn"/"kube-root-ca.crt" Dec 10 20:30:32 crc kubenswrapper[4828]: I1210 20:30:32.148694 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qpwwn/must-gather-n5s68"] Dec 10 20:30:32 crc kubenswrapper[4828]: I1210 20:30:32.185105 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e4454add-ea32-4989-a3ef-700e057eed1a-must-gather-output\") pod \"must-gather-n5s68\" (UID: \"e4454add-ea32-4989-a3ef-700e057eed1a\") " pod="openshift-must-gather-qpwwn/must-gather-n5s68" Dec 10 20:30:32 crc kubenswrapper[4828]: I1210 20:30:32.186407 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9pff\" (UniqueName: \"kubernetes.io/projected/e4454add-ea32-4989-a3ef-700e057eed1a-kube-api-access-k9pff\") pod \"must-gather-n5s68\" (UID: \"e4454add-ea32-4989-a3ef-700e057eed1a\") " pod="openshift-must-gather-qpwwn/must-gather-n5s68" Dec 10 20:30:32 crc kubenswrapper[4828]: I1210 20:30:32.290060 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e4454add-ea32-4989-a3ef-700e057eed1a-must-gather-output\") pod \"must-gather-n5s68\" (UID: \"e4454add-ea32-4989-a3ef-700e057eed1a\") " pod="openshift-must-gather-qpwwn/must-gather-n5s68" Dec 10 20:30:32 crc kubenswrapper[4828]: I1210 20:30:32.290148 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9pff\" (UniqueName: \"kubernetes.io/projected/e4454add-ea32-4989-a3ef-700e057eed1a-kube-api-access-k9pff\") pod \"must-gather-n5s68\" (UID: \"e4454add-ea32-4989-a3ef-700e057eed1a\") " pod="openshift-must-gather-qpwwn/must-gather-n5s68" Dec 10 20:30:32 crc kubenswrapper[4828]: I1210 20:30:32.290990 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e4454add-ea32-4989-a3ef-700e057eed1a-must-gather-output\") pod \"must-gather-n5s68\" (UID: \"e4454add-ea32-4989-a3ef-700e057eed1a\") " pod="openshift-must-gather-qpwwn/must-gather-n5s68" Dec 10 20:30:32 crc kubenswrapper[4828]: I1210 20:30:32.380482 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9pff\" (UniqueName: \"kubernetes.io/projected/e4454add-ea32-4989-a3ef-700e057eed1a-kube-api-access-k9pff\") pod \"must-gather-n5s68\" (UID: \"e4454add-ea32-4989-a3ef-700e057eed1a\") " pod="openshift-must-gather-qpwwn/must-gather-n5s68" Dec 10 20:30:32 crc kubenswrapper[4828]: I1210 20:30:32.460750 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qpwwn/must-gather-n5s68" Dec 10 20:30:32 crc kubenswrapper[4828]: I1210 20:30:32.957380 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qpwwn/must-gather-n5s68"] Dec 10 20:30:32 crc kubenswrapper[4828]: W1210 20:30:32.973623 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4454add_ea32_4989_a3ef_700e057eed1a.slice/crio-9f65c8fba51e6209fc461637b324586371e0a89146023c40fa49df027beceef1 WatchSource:0}: Error finding container 9f65c8fba51e6209fc461637b324586371e0a89146023c40fa49df027beceef1: Status 404 returned error can't find the container with id 9f65c8fba51e6209fc461637b324586371e0a89146023c40fa49df027beceef1 Dec 10 20:30:33 crc kubenswrapper[4828]: I1210 20:30:33.885435 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qpwwn/must-gather-n5s68" event={"ID":"e4454add-ea32-4989-a3ef-700e057eed1a","Type":"ContainerStarted","Data":"9f65c8fba51e6209fc461637b324586371e0a89146023c40fa49df027beceef1"} Dec 10 20:30:41 crc kubenswrapper[4828]: I1210 20:30:41.993198 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qpwwn/must-gather-n5s68" event={"ID":"e4454add-ea32-4989-a3ef-700e057eed1a","Type":"ContainerStarted","Data":"324ffea16ebb8fc04036ee280c3a99820b4d0b87426ffd21adb809e90a4d7f72"} Dec 10 20:30:41 crc kubenswrapper[4828]: I1210 20:30:41.994810 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qpwwn/must-gather-n5s68" event={"ID":"e4454add-ea32-4989-a3ef-700e057eed1a","Type":"ContainerStarted","Data":"0a1fe51129829cad3c79669a585e9268fbd56911b8d54f0f5e4d174b4743f60c"} Dec 10 20:30:42 crc kubenswrapper[4828]: I1210 20:30:42.021407 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qpwwn/must-gather-n5s68" podStartSLOduration=2.179552934 podStartE2EDuration="10.021385328s" podCreationTimestamp="2025-12-10 20:30:32 +0000 UTC" firstStartedPulling="2025-12-10 20:30:32.980109561 +0000 UTC m=+5713.490720566" lastFinishedPulling="2025-12-10 20:30:40.821941945 +0000 UTC m=+5721.332552960" observedRunningTime="2025-12-10 20:30:42.011992157 +0000 UTC m=+5722.522603162" watchObservedRunningTime="2025-12-10 20:30:42.021385328 +0000 UTC m=+5722.531996333" Dec 10 20:30:42 crc kubenswrapper[4828]: I1210 20:30:42.789674 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:30:42 crc kubenswrapper[4828]: E1210 20:30:42.789944 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:30:44 crc kubenswrapper[4828]: I1210 20:30:44.893653 4828 scope.go:117] "RemoveContainer" containerID="1960712b3a4058cacb3039abfd7810ef09e544bd8f9ef480338e2450963e71e6" Dec 10 20:30:46 crc kubenswrapper[4828]: I1210 20:30:46.195080 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qpwwn/crc-debug-992bq"] Dec 10 20:30:46 crc kubenswrapper[4828]: I1210 20:30:46.197710 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qpwwn/crc-debug-992bq" Dec 10 20:30:46 crc kubenswrapper[4828]: I1210 20:30:46.205742 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4ca1e93f-59cf-4103-a602-dc1925ed2378-host\") pod \"crc-debug-992bq\" (UID: \"4ca1e93f-59cf-4103-a602-dc1925ed2378\") " pod="openshift-must-gather-qpwwn/crc-debug-992bq" Dec 10 20:30:46 crc kubenswrapper[4828]: I1210 20:30:46.205831 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8lb9\" (UniqueName: \"kubernetes.io/projected/4ca1e93f-59cf-4103-a602-dc1925ed2378-kube-api-access-g8lb9\") pod \"crc-debug-992bq\" (UID: \"4ca1e93f-59cf-4103-a602-dc1925ed2378\") " pod="openshift-must-gather-qpwwn/crc-debug-992bq" Dec 10 20:30:46 crc kubenswrapper[4828]: I1210 20:30:46.308610 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4ca1e93f-59cf-4103-a602-dc1925ed2378-host\") pod \"crc-debug-992bq\" (UID: \"4ca1e93f-59cf-4103-a602-dc1925ed2378\") " pod="openshift-must-gather-qpwwn/crc-debug-992bq" Dec 10 20:30:46 crc kubenswrapper[4828]: I1210 20:30:46.308984 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8lb9\" (UniqueName: \"kubernetes.io/projected/4ca1e93f-59cf-4103-a602-dc1925ed2378-kube-api-access-g8lb9\") pod \"crc-debug-992bq\" (UID: \"4ca1e93f-59cf-4103-a602-dc1925ed2378\") " pod="openshift-must-gather-qpwwn/crc-debug-992bq" Dec 10 20:30:46 crc kubenswrapper[4828]: I1210 20:30:46.309419 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4ca1e93f-59cf-4103-a602-dc1925ed2378-host\") pod \"crc-debug-992bq\" (UID: \"4ca1e93f-59cf-4103-a602-dc1925ed2378\") " pod="openshift-must-gather-qpwwn/crc-debug-992bq" Dec 10 20:30:46 crc kubenswrapper[4828]: I1210 20:30:46.329434 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8lb9\" (UniqueName: \"kubernetes.io/projected/4ca1e93f-59cf-4103-a602-dc1925ed2378-kube-api-access-g8lb9\") pod \"crc-debug-992bq\" (UID: \"4ca1e93f-59cf-4103-a602-dc1925ed2378\") " pod="openshift-must-gather-qpwwn/crc-debug-992bq" Dec 10 20:30:46 crc kubenswrapper[4828]: I1210 20:30:46.517038 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qpwwn/crc-debug-992bq" Dec 10 20:30:47 crc kubenswrapper[4828]: I1210 20:30:47.047371 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qpwwn/crc-debug-992bq" event={"ID":"4ca1e93f-59cf-4103-a602-dc1925ed2378","Type":"ContainerStarted","Data":"bcacefa3d6c239e3907f42214070650f3b099ea3adbe9d2ed9b313b43aa3ebac"} Dec 10 20:30:56 crc kubenswrapper[4828]: I1210 20:30:56.788639 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:30:56 crc kubenswrapper[4828]: E1210 20:30:56.789670 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:30:57 crc kubenswrapper[4828]: I1210 20:30:57.193932 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qpwwn/crc-debug-992bq" event={"ID":"4ca1e93f-59cf-4103-a602-dc1925ed2378","Type":"ContainerStarted","Data":"cd93cc78c51aae8e287b9abfebc85e1f6c9880808a1dbb8e44e760a74eb3aa12"} Dec 10 20:30:57 crc kubenswrapper[4828]: I1210 20:30:57.214338 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qpwwn/crc-debug-992bq" podStartSLOduration=1.256633852 podStartE2EDuration="11.214319989s" podCreationTimestamp="2025-12-10 20:30:46 +0000 UTC" firstStartedPulling="2025-12-10 20:30:46.596835618 +0000 UTC m=+5727.107446623" lastFinishedPulling="2025-12-10 20:30:56.554521765 +0000 UTC m=+5737.065132760" observedRunningTime="2025-12-10 20:30:57.210130987 +0000 UTC m=+5737.720741992" watchObservedRunningTime="2025-12-10 20:30:57.214319989 +0000 UTC m=+5737.724930994" Dec 10 20:31:07 crc kubenswrapper[4828]: I1210 20:31:07.790208 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:31:07 crc kubenswrapper[4828]: E1210 20:31:07.793030 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:31:22 crc kubenswrapper[4828]: I1210 20:31:22.789870 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:31:22 crc kubenswrapper[4828]: E1210 20:31:22.790934 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:31:37 crc kubenswrapper[4828]: I1210 20:31:37.789216 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:31:37 crc kubenswrapper[4828]: E1210 20:31:37.790128 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:31:50 crc kubenswrapper[4828]: I1210 20:31:50.788936 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:31:50 crc kubenswrapper[4828]: E1210 20:31:50.790029 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:31:50 crc kubenswrapper[4828]: I1210 20:31:50.800068 4828 generic.go:334] "Generic (PLEG): container finished" podID="4ca1e93f-59cf-4103-a602-dc1925ed2378" containerID="cd93cc78c51aae8e287b9abfebc85e1f6c9880808a1dbb8e44e760a74eb3aa12" exitCode=0 Dec 10 20:31:50 crc kubenswrapper[4828]: I1210 20:31:50.800123 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qpwwn/crc-debug-992bq" event={"ID":"4ca1e93f-59cf-4103-a602-dc1925ed2378","Type":"ContainerDied","Data":"cd93cc78c51aae8e287b9abfebc85e1f6c9880808a1dbb8e44e760a74eb3aa12"} Dec 10 20:31:51 crc kubenswrapper[4828]: I1210 20:31:51.937883 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qpwwn/crc-debug-992bq" Dec 10 20:31:51 crc kubenswrapper[4828]: I1210 20:31:51.974542 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qpwwn/crc-debug-992bq"] Dec 10 20:31:51 crc kubenswrapper[4828]: I1210 20:31:51.986055 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qpwwn/crc-debug-992bq"] Dec 10 20:31:52 crc kubenswrapper[4828]: I1210 20:31:52.094394 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4ca1e93f-59cf-4103-a602-dc1925ed2378-host\") pod \"4ca1e93f-59cf-4103-a602-dc1925ed2378\" (UID: \"4ca1e93f-59cf-4103-a602-dc1925ed2378\") " Dec 10 20:31:52 crc kubenswrapper[4828]: I1210 20:31:52.094602 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ca1e93f-59cf-4103-a602-dc1925ed2378-host" (OuterVolumeSpecName: "host") pod "4ca1e93f-59cf-4103-a602-dc1925ed2378" (UID: "4ca1e93f-59cf-4103-a602-dc1925ed2378"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 20:31:52 crc kubenswrapper[4828]: I1210 20:31:52.094650 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8lb9\" (UniqueName: \"kubernetes.io/projected/4ca1e93f-59cf-4103-a602-dc1925ed2378-kube-api-access-g8lb9\") pod \"4ca1e93f-59cf-4103-a602-dc1925ed2378\" (UID: \"4ca1e93f-59cf-4103-a602-dc1925ed2378\") " Dec 10 20:31:52 crc kubenswrapper[4828]: I1210 20:31:52.095236 4828 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4ca1e93f-59cf-4103-a602-dc1925ed2378-host\") on node \"crc\" DevicePath \"\"" Dec 10 20:31:52 crc kubenswrapper[4828]: I1210 20:31:52.100489 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ca1e93f-59cf-4103-a602-dc1925ed2378-kube-api-access-g8lb9" (OuterVolumeSpecName: "kube-api-access-g8lb9") pod "4ca1e93f-59cf-4103-a602-dc1925ed2378" (UID: "4ca1e93f-59cf-4103-a602-dc1925ed2378"). InnerVolumeSpecName "kube-api-access-g8lb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:31:52 crc kubenswrapper[4828]: I1210 20:31:52.197420 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8lb9\" (UniqueName: \"kubernetes.io/projected/4ca1e93f-59cf-4103-a602-dc1925ed2378-kube-api-access-g8lb9\") on node \"crc\" DevicePath \"\"" Dec 10 20:31:52 crc kubenswrapper[4828]: I1210 20:31:52.821672 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcacefa3d6c239e3907f42214070650f3b099ea3adbe9d2ed9b313b43aa3ebac" Dec 10 20:31:52 crc kubenswrapper[4828]: I1210 20:31:52.821737 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qpwwn/crc-debug-992bq" Dec 10 20:31:53 crc kubenswrapper[4828]: I1210 20:31:53.138904 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qpwwn/crc-debug-nxnpq"] Dec 10 20:31:53 crc kubenswrapper[4828]: E1210 20:31:53.139370 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ca1e93f-59cf-4103-a602-dc1925ed2378" containerName="container-00" Dec 10 20:31:53 crc kubenswrapper[4828]: I1210 20:31:53.139385 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca1e93f-59cf-4103-a602-dc1925ed2378" containerName="container-00" Dec 10 20:31:53 crc kubenswrapper[4828]: I1210 20:31:53.139718 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ca1e93f-59cf-4103-a602-dc1925ed2378" containerName="container-00" Dec 10 20:31:53 crc kubenswrapper[4828]: I1210 20:31:53.140491 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qpwwn/crc-debug-nxnpq" Dec 10 20:31:53 crc kubenswrapper[4828]: I1210 20:31:53.217767 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9ebee0b-4ee8-4341-94b7-666c41ac002b-host\") pod \"crc-debug-nxnpq\" (UID: \"b9ebee0b-4ee8-4341-94b7-666c41ac002b\") " pod="openshift-must-gather-qpwwn/crc-debug-nxnpq" Dec 10 20:31:53 crc kubenswrapper[4828]: I1210 20:31:53.218130 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ws9s\" (UniqueName: \"kubernetes.io/projected/b9ebee0b-4ee8-4341-94b7-666c41ac002b-kube-api-access-6ws9s\") pod \"crc-debug-nxnpq\" (UID: \"b9ebee0b-4ee8-4341-94b7-666c41ac002b\") " pod="openshift-must-gather-qpwwn/crc-debug-nxnpq" Dec 10 20:31:53 crc kubenswrapper[4828]: I1210 20:31:53.320642 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9ebee0b-4ee8-4341-94b7-666c41ac002b-host\") pod \"crc-debug-nxnpq\" (UID: \"b9ebee0b-4ee8-4341-94b7-666c41ac002b\") " pod="openshift-must-gather-qpwwn/crc-debug-nxnpq" Dec 10 20:31:53 crc kubenswrapper[4828]: I1210 20:31:53.320997 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ws9s\" (UniqueName: \"kubernetes.io/projected/b9ebee0b-4ee8-4341-94b7-666c41ac002b-kube-api-access-6ws9s\") pod \"crc-debug-nxnpq\" (UID: \"b9ebee0b-4ee8-4341-94b7-666c41ac002b\") " pod="openshift-must-gather-qpwwn/crc-debug-nxnpq" Dec 10 20:31:53 crc kubenswrapper[4828]: I1210 20:31:53.320779 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9ebee0b-4ee8-4341-94b7-666c41ac002b-host\") pod \"crc-debug-nxnpq\" (UID: \"b9ebee0b-4ee8-4341-94b7-666c41ac002b\") " pod="openshift-must-gather-qpwwn/crc-debug-nxnpq" Dec 10 20:31:53 crc kubenswrapper[4828]: I1210 20:31:53.341312 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ws9s\" (UniqueName: \"kubernetes.io/projected/b9ebee0b-4ee8-4341-94b7-666c41ac002b-kube-api-access-6ws9s\") pod \"crc-debug-nxnpq\" (UID: \"b9ebee0b-4ee8-4341-94b7-666c41ac002b\") " pod="openshift-must-gather-qpwwn/crc-debug-nxnpq" Dec 10 20:31:53 crc kubenswrapper[4828]: I1210 20:31:53.461530 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qpwwn/crc-debug-nxnpq" Dec 10 20:31:53 crc kubenswrapper[4828]: I1210 20:31:53.802597 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ca1e93f-59cf-4103-a602-dc1925ed2378" path="/var/lib/kubelet/pods/4ca1e93f-59cf-4103-a602-dc1925ed2378/volumes" Dec 10 20:31:53 crc kubenswrapper[4828]: I1210 20:31:53.839186 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qpwwn/crc-debug-nxnpq" event={"ID":"b9ebee0b-4ee8-4341-94b7-666c41ac002b","Type":"ContainerStarted","Data":"d7ab96cf13291f0ca1e526f6b65a4f4ae38ee562512b6d4e6b49de28c66d1ccf"} Dec 10 20:31:53 crc kubenswrapper[4828]: I1210 20:31:53.839262 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qpwwn/crc-debug-nxnpq" event={"ID":"b9ebee0b-4ee8-4341-94b7-666c41ac002b","Type":"ContainerStarted","Data":"0ae91675cee291a4917232d23f71772fed78b8fc1a8d5dfe8f4f3d7e8ad311f5"} Dec 10 20:31:53 crc kubenswrapper[4828]: I1210 20:31:53.859555 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qpwwn/crc-debug-nxnpq" podStartSLOduration=0.859530236 podStartE2EDuration="859.530236ms" podCreationTimestamp="2025-12-10 20:31:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-10 20:31:53.852690973 +0000 UTC m=+5794.363301978" watchObservedRunningTime="2025-12-10 20:31:53.859530236 +0000 UTC m=+5794.370141241" Dec 10 20:31:54 crc kubenswrapper[4828]: I1210 20:31:54.853008 4828 generic.go:334] "Generic (PLEG): container finished" podID="b9ebee0b-4ee8-4341-94b7-666c41ac002b" containerID="d7ab96cf13291f0ca1e526f6b65a4f4ae38ee562512b6d4e6b49de28c66d1ccf" exitCode=0 Dec 10 20:31:54 crc kubenswrapper[4828]: I1210 20:31:54.853307 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qpwwn/crc-debug-nxnpq" event={"ID":"b9ebee0b-4ee8-4341-94b7-666c41ac002b","Type":"ContainerDied","Data":"d7ab96cf13291f0ca1e526f6b65a4f4ae38ee562512b6d4e6b49de28c66d1ccf"} Dec 10 20:31:55 crc kubenswrapper[4828]: I1210 20:31:55.997715 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qpwwn/crc-debug-nxnpq" Dec 10 20:31:56 crc kubenswrapper[4828]: I1210 20:31:56.110489 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ws9s\" (UniqueName: \"kubernetes.io/projected/b9ebee0b-4ee8-4341-94b7-666c41ac002b-kube-api-access-6ws9s\") pod \"b9ebee0b-4ee8-4341-94b7-666c41ac002b\" (UID: \"b9ebee0b-4ee8-4341-94b7-666c41ac002b\") " Dec 10 20:31:56 crc kubenswrapper[4828]: I1210 20:31:56.111001 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9ebee0b-4ee8-4341-94b7-666c41ac002b-host\") pod \"b9ebee0b-4ee8-4341-94b7-666c41ac002b\" (UID: \"b9ebee0b-4ee8-4341-94b7-666c41ac002b\") " Dec 10 20:31:56 crc kubenswrapper[4828]: I1210 20:31:56.111129 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9ebee0b-4ee8-4341-94b7-666c41ac002b-host" (OuterVolumeSpecName: "host") pod "b9ebee0b-4ee8-4341-94b7-666c41ac002b" (UID: "b9ebee0b-4ee8-4341-94b7-666c41ac002b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 20:31:56 crc kubenswrapper[4828]: I1210 20:31:56.111936 4828 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9ebee0b-4ee8-4341-94b7-666c41ac002b-host\") on node \"crc\" DevicePath \"\"" Dec 10 20:31:56 crc kubenswrapper[4828]: I1210 20:31:56.117975 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9ebee0b-4ee8-4341-94b7-666c41ac002b-kube-api-access-6ws9s" (OuterVolumeSpecName: "kube-api-access-6ws9s") pod "b9ebee0b-4ee8-4341-94b7-666c41ac002b" (UID: "b9ebee0b-4ee8-4341-94b7-666c41ac002b"). InnerVolumeSpecName "kube-api-access-6ws9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:31:56 crc kubenswrapper[4828]: I1210 20:31:56.213861 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ws9s\" (UniqueName: \"kubernetes.io/projected/b9ebee0b-4ee8-4341-94b7-666c41ac002b-kube-api-access-6ws9s\") on node \"crc\" DevicePath \"\"" Dec 10 20:31:56 crc kubenswrapper[4828]: I1210 20:31:56.272442 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qpwwn/crc-debug-nxnpq"] Dec 10 20:31:56 crc kubenswrapper[4828]: I1210 20:31:56.282347 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qpwwn/crc-debug-nxnpq"] Dec 10 20:31:56 crc kubenswrapper[4828]: I1210 20:31:56.875537 4828 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ae91675cee291a4917232d23f71772fed78b8fc1a8d5dfe8f4f3d7e8ad311f5" Dec 10 20:31:56 crc kubenswrapper[4828]: I1210 20:31:56.875615 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qpwwn/crc-debug-nxnpq" Dec 10 20:31:57 crc kubenswrapper[4828]: I1210 20:31:57.437766 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qpwwn/crc-debug-589zb"] Dec 10 20:31:57 crc kubenswrapper[4828]: E1210 20:31:57.438347 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9ebee0b-4ee8-4341-94b7-666c41ac002b" containerName="container-00" Dec 10 20:31:57 crc kubenswrapper[4828]: I1210 20:31:57.438363 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9ebee0b-4ee8-4341-94b7-666c41ac002b" containerName="container-00" Dec 10 20:31:57 crc kubenswrapper[4828]: I1210 20:31:57.438596 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9ebee0b-4ee8-4341-94b7-666c41ac002b" containerName="container-00" Dec 10 20:31:57 crc kubenswrapper[4828]: I1210 20:31:57.439451 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qpwwn/crc-debug-589zb" Dec 10 20:31:57 crc kubenswrapper[4828]: I1210 20:31:57.542416 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kdzs\" (UniqueName: \"kubernetes.io/projected/54ac3720-5d29-4c80-885a-33ce3a73bdf2-kube-api-access-5kdzs\") pod \"crc-debug-589zb\" (UID: \"54ac3720-5d29-4c80-885a-33ce3a73bdf2\") " pod="openshift-must-gather-qpwwn/crc-debug-589zb" Dec 10 20:31:57 crc kubenswrapper[4828]: I1210 20:31:57.542703 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/54ac3720-5d29-4c80-885a-33ce3a73bdf2-host\") pod \"crc-debug-589zb\" (UID: \"54ac3720-5d29-4c80-885a-33ce3a73bdf2\") " pod="openshift-must-gather-qpwwn/crc-debug-589zb" Dec 10 20:31:57 crc kubenswrapper[4828]: I1210 20:31:57.644851 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kdzs\" (UniqueName: \"kubernetes.io/projected/54ac3720-5d29-4c80-885a-33ce3a73bdf2-kube-api-access-5kdzs\") pod \"crc-debug-589zb\" (UID: \"54ac3720-5d29-4c80-885a-33ce3a73bdf2\") " pod="openshift-must-gather-qpwwn/crc-debug-589zb" Dec 10 20:31:57 crc kubenswrapper[4828]: I1210 20:31:57.645172 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/54ac3720-5d29-4c80-885a-33ce3a73bdf2-host\") pod \"crc-debug-589zb\" (UID: \"54ac3720-5d29-4c80-885a-33ce3a73bdf2\") " pod="openshift-must-gather-qpwwn/crc-debug-589zb" Dec 10 20:31:57 crc kubenswrapper[4828]: I1210 20:31:57.645269 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/54ac3720-5d29-4c80-885a-33ce3a73bdf2-host\") pod \"crc-debug-589zb\" (UID: \"54ac3720-5d29-4c80-885a-33ce3a73bdf2\") " pod="openshift-must-gather-qpwwn/crc-debug-589zb" Dec 10 20:31:57 crc kubenswrapper[4828]: I1210 20:31:57.664837 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kdzs\" (UniqueName: \"kubernetes.io/projected/54ac3720-5d29-4c80-885a-33ce3a73bdf2-kube-api-access-5kdzs\") pod \"crc-debug-589zb\" (UID: \"54ac3720-5d29-4c80-885a-33ce3a73bdf2\") " pod="openshift-must-gather-qpwwn/crc-debug-589zb" Dec 10 20:31:57 crc kubenswrapper[4828]: I1210 20:31:57.781553 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qpwwn/crc-debug-589zb" Dec 10 20:31:57 crc kubenswrapper[4828]: I1210 20:31:57.805741 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9ebee0b-4ee8-4341-94b7-666c41ac002b" path="/var/lib/kubelet/pods/b9ebee0b-4ee8-4341-94b7-666c41ac002b/volumes" Dec 10 20:31:57 crc kubenswrapper[4828]: W1210 20:31:57.812887 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54ac3720_5d29_4c80_885a_33ce3a73bdf2.slice/crio-2b36b8801deb0bd597f43a4d9afe12e761849ef9524298e360f84f3a9e6cd37b WatchSource:0}: Error finding container 2b36b8801deb0bd597f43a4d9afe12e761849ef9524298e360f84f3a9e6cd37b: Status 404 returned error can't find the container with id 2b36b8801deb0bd597f43a4d9afe12e761849ef9524298e360f84f3a9e6cd37b Dec 10 20:31:57 crc kubenswrapper[4828]: I1210 20:31:57.888665 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qpwwn/crc-debug-589zb" event={"ID":"54ac3720-5d29-4c80-885a-33ce3a73bdf2","Type":"ContainerStarted","Data":"2b36b8801deb0bd597f43a4d9afe12e761849ef9524298e360f84f3a9e6cd37b"} Dec 10 20:31:58 crc kubenswrapper[4828]: I1210 20:31:58.902667 4828 generic.go:334] "Generic (PLEG): container finished" podID="54ac3720-5d29-4c80-885a-33ce3a73bdf2" containerID="1eeab2398bdc70cd5a1b2127799e04b9a80c13ed4ad81f814bd1629791bc0b4f" exitCode=0 Dec 10 20:31:58 crc kubenswrapper[4828]: I1210 20:31:58.902734 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qpwwn/crc-debug-589zb" event={"ID":"54ac3720-5d29-4c80-885a-33ce3a73bdf2","Type":"ContainerDied","Data":"1eeab2398bdc70cd5a1b2127799e04b9a80c13ed4ad81f814bd1629791bc0b4f"} Dec 10 20:31:58 crc kubenswrapper[4828]: I1210 20:31:58.947331 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qpwwn/crc-debug-589zb"] Dec 10 20:31:58 crc kubenswrapper[4828]: I1210 20:31:58.957744 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qpwwn/crc-debug-589zb"] Dec 10 20:32:00 crc kubenswrapper[4828]: I1210 20:32:00.045358 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qpwwn/crc-debug-589zb" Dec 10 20:32:00 crc kubenswrapper[4828]: I1210 20:32:00.205631 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/54ac3720-5d29-4c80-885a-33ce3a73bdf2-host\") pod \"54ac3720-5d29-4c80-885a-33ce3a73bdf2\" (UID: \"54ac3720-5d29-4c80-885a-33ce3a73bdf2\") " Dec 10 20:32:00 crc kubenswrapper[4828]: I1210 20:32:00.205774 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kdzs\" (UniqueName: \"kubernetes.io/projected/54ac3720-5d29-4c80-885a-33ce3a73bdf2-kube-api-access-5kdzs\") pod \"54ac3720-5d29-4c80-885a-33ce3a73bdf2\" (UID: \"54ac3720-5d29-4c80-885a-33ce3a73bdf2\") " Dec 10 20:32:00 crc kubenswrapper[4828]: I1210 20:32:00.205764 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/54ac3720-5d29-4c80-885a-33ce3a73bdf2-host" (OuterVolumeSpecName: "host") pod "54ac3720-5d29-4c80-885a-33ce3a73bdf2" (UID: "54ac3720-5d29-4c80-885a-33ce3a73bdf2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 10 20:32:00 crc kubenswrapper[4828]: I1210 20:32:00.206666 4828 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/54ac3720-5d29-4c80-885a-33ce3a73bdf2-host\") on node \"crc\" DevicePath \"\"" Dec 10 20:32:00 crc kubenswrapper[4828]: I1210 20:32:00.216054 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54ac3720-5d29-4c80-885a-33ce3a73bdf2-kube-api-access-5kdzs" (OuterVolumeSpecName: "kube-api-access-5kdzs") pod "54ac3720-5d29-4c80-885a-33ce3a73bdf2" (UID: "54ac3720-5d29-4c80-885a-33ce3a73bdf2"). InnerVolumeSpecName "kube-api-access-5kdzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:32:00 crc kubenswrapper[4828]: I1210 20:32:00.308962 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kdzs\" (UniqueName: \"kubernetes.io/projected/54ac3720-5d29-4c80-885a-33ce3a73bdf2-kube-api-access-5kdzs\") on node \"crc\" DevicePath \"\"" Dec 10 20:32:00 crc kubenswrapper[4828]: I1210 20:32:00.928784 4828 scope.go:117] "RemoveContainer" containerID="1eeab2398bdc70cd5a1b2127799e04b9a80c13ed4ad81f814bd1629791bc0b4f" Dec 10 20:32:00 crc kubenswrapper[4828]: I1210 20:32:00.928846 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qpwwn/crc-debug-589zb" Dec 10 20:32:01 crc kubenswrapper[4828]: I1210 20:32:01.806692 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54ac3720-5d29-4c80-885a-33ce3a73bdf2" path="/var/lib/kubelet/pods/54ac3720-5d29-4c80-885a-33ce3a73bdf2/volumes" Dec 10 20:32:03 crc kubenswrapper[4828]: I1210 20:32:03.789424 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:32:03 crc kubenswrapper[4828]: E1210 20:32:03.790281 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:32:16 crc kubenswrapper[4828]: I1210 20:32:16.789539 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:32:16 crc kubenswrapper[4828]: E1210 20:32:16.790436 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:32:26 crc kubenswrapper[4828]: I1210 20:32:26.274149 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_683b1cf4-b6dc-4bf5-a537-6461533fc1ef/aodh-api/0.log" Dec 10 20:32:26 crc kubenswrapper[4828]: I1210 20:32:26.325959 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_683b1cf4-b6dc-4bf5-a537-6461533fc1ef/aodh-evaluator/0.log" Dec 10 20:32:26 crc kubenswrapper[4828]: I1210 20:32:26.404195 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_683b1cf4-b6dc-4bf5-a537-6461533fc1ef/aodh-listener/0.log" Dec 10 20:32:26 crc kubenswrapper[4828]: I1210 20:32:26.478518 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_683b1cf4-b6dc-4bf5-a537-6461533fc1ef/aodh-notifier/0.log" Dec 10 20:32:26 crc kubenswrapper[4828]: I1210 20:32:26.590745 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6598f5c59d-btf5k_d425401e-33bc-4dc1-adce-44622ce940df/barbican-api/0.log" Dec 10 20:32:26 crc kubenswrapper[4828]: I1210 20:32:26.621462 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6598f5c59d-btf5k_d425401e-33bc-4dc1-adce-44622ce940df/barbican-api-log/0.log" Dec 10 20:32:26 crc kubenswrapper[4828]: I1210 20:32:26.799947 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-85965bf486-fgbxr_1bc1ebd1-1426-4def-a07d-5b0b561f2082/barbican-keystone-listener/0.log" Dec 10 20:32:26 crc kubenswrapper[4828]: I1210 20:32:26.940245 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-85965bf486-fgbxr_1bc1ebd1-1426-4def-a07d-5b0b561f2082/barbican-keystone-listener-log/0.log" Dec 10 20:32:27 crc kubenswrapper[4828]: I1210 20:32:27.031304 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-69884f6747-2l55b_7ee9fcef-2371-4614-b67f-6cef46b20c6f/barbican-worker/0.log" Dec 10 20:32:27 crc kubenswrapper[4828]: I1210 20:32:27.076226 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-69884f6747-2l55b_7ee9fcef-2371-4614-b67f-6cef46b20c6f/barbican-worker-log/0.log" Dec 10 20:32:27 crc kubenswrapper[4828]: I1210 20:32:27.232531 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-dwr4m_c937bc31-7116-4ab2-bd4e-c5e2bfd10e69/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:27 crc kubenswrapper[4828]: I1210 20:32:27.422698 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_cd0c6bc2-1b33-4c82-ae2f-9ea265564b82/ceilometer-central-agent/0.log" Dec 10 20:32:27 crc kubenswrapper[4828]: I1210 20:32:27.464056 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_cd0c6bc2-1b33-4c82-ae2f-9ea265564b82/proxy-httpd/0.log" Dec 10 20:32:27 crc kubenswrapper[4828]: I1210 20:32:27.501438 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_cd0c6bc2-1b33-4c82-ae2f-9ea265564b82/ceilometer-notification-agent/0.log" Dec 10 20:32:27 crc kubenswrapper[4828]: I1210 20:32:27.570438 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_cd0c6bc2-1b33-4c82-ae2f-9ea265564b82/sg-core/0.log" Dec 10 20:32:27 crc kubenswrapper[4828]: I1210 20:32:27.751436 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_71db2c7f-94ac-455b-9142-c63bb53ed942/cinder-api/0.log" Dec 10 20:32:27 crc kubenswrapper[4828]: I1210 20:32:27.770617 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_71db2c7f-94ac-455b-9142-c63bb53ed942/cinder-api-log/0.log" Dec 10 20:32:28 crc kubenswrapper[4828]: I1210 20:32:28.073648 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8f65c47b-68bb-4c64-b0e8-94507fb828e5/cinder-scheduler/0.log" Dec 10 20:32:28 crc kubenswrapper[4828]: I1210 20:32:28.136306 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_8f65c47b-68bb-4c64-b0e8-94507fb828e5/probe/0.log" Dec 10 20:32:28 crc kubenswrapper[4828]: I1210 20:32:28.250541 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-g4ghf_88eb8a6c-2e65-42b9-bda9-89c45196fd75/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:28 crc kubenswrapper[4828]: I1210 20:32:28.394959 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-xd599_b3f83864-18ee-4493-b420-a536371c509b/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:28 crc kubenswrapper[4828]: I1210 20:32:28.495756 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-9xsbz_b460f485-38dd-48c1-81d5-288cd81f392b/init/0.log" Dec 10 20:32:28 crc kubenswrapper[4828]: I1210 20:32:28.893930 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-9xsbz_b460f485-38dd-48c1-81d5-288cd81f392b/init/0.log" Dec 10 20:32:28 crc kubenswrapper[4828]: I1210 20:32:28.977547 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-qjjsq_98d9383c-0b06-4a8d-8d4c-1832493c3c5f/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:29 crc kubenswrapper[4828]: I1210 20:32:29.010829 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-9xsbz_b460f485-38dd-48c1-81d5-288cd81f392b/dnsmasq-dns/0.log" Dec 10 20:32:29 crc kubenswrapper[4828]: I1210 20:32:29.206265 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2fffc8e4-c48c-4f89-bc5d-143c3d96a557/glance-httpd/0.log" Dec 10 20:32:29 crc kubenswrapper[4828]: I1210 20:32:29.249709 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2fffc8e4-c48c-4f89-bc5d-143c3d96a557/glance-log/0.log" Dec 10 20:32:29 crc kubenswrapper[4828]: I1210 20:32:29.416156 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3c08f474-3395-463f-ac2f-559aa97bb021/glance-log/0.log" Dec 10 20:32:29 crc kubenswrapper[4828]: I1210 20:32:29.449888 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3c08f474-3395-463f-ac2f-559aa97bb021/glance-httpd/0.log" Dec 10 20:32:29 crc kubenswrapper[4828]: I1210 20:32:29.799478 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:32:29 crc kubenswrapper[4828]: E1210 20:32:29.799980 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:32:30 crc kubenswrapper[4828]: I1210 20:32:30.069785 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-55d5c67ddc-vcr2w_7dd195e2-5f2e-4472-a481-bfb485dcfc74/heat-engine/0.log" Dec 10 20:32:30 crc kubenswrapper[4828]: I1210 20:32:30.219342 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-5859bd5774-r9zl7_961e506a-98a1-48e1-afd5-c4326af0bdb3/heat-api/0.log" Dec 10 20:32:30 crc kubenswrapper[4828]: I1210 20:32:30.318774 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-wmhjv_5726d3d2-c797-4fca-aa6c-8aec98c6061b/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:30 crc kubenswrapper[4828]: I1210 20:32:30.340234 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-cb5d96886-n5cp9_d1887db4-171d-4bd1-8337-4a8c72e9e93d/heat-cfnapi/0.log" Dec 10 20:32:30 crc kubenswrapper[4828]: I1210 20:32:30.460643 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-snmlm_2bb52b6f-bc4a-411d-be87-eb45eccbeb4a/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:30 crc kubenswrapper[4828]: I1210 20:32:30.740161 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29423281-rkwnp_40346cd5-252e-4737-a7ff-8dacb87c3a7f/keystone-cron/0.log" Dec 10 20:32:30 crc kubenswrapper[4828]: I1210 20:32:30.908992 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_04203c31-30dc-48d1-bdb9-1a8eaf8da709/kube-state-metrics/0.log" Dec 10 20:32:31 crc kubenswrapper[4828]: I1210 20:32:31.017971 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6f465c7bf6-5f8bw_aaa7ff68-b5a6-4f74-9879-6a51c6016578/keystone-api/0.log" Dec 10 20:32:31 crc kubenswrapper[4828]: I1210 20:32:31.029315 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-qjk2w_791b45cd-a74c-493b-b33b-e97c642318dc/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:31 crc kubenswrapper[4828]: I1210 20:32:31.167998 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-2lb42_94949893-8bdc-4a05-b105-8759681782ae/logging-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:31 crc kubenswrapper[4828]: I1210 20:32:31.358354 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_198e58a9-696b-4857-a6fe-f31cf31d61d7/mysqld-exporter/0.log" Dec 10 20:32:31 crc kubenswrapper[4828]: I1210 20:32:31.687330 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-hbgkj_c502be9a-c873-4b9b-97e8-e6461052d252/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:31 crc kubenswrapper[4828]: I1210 20:32:31.689117 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-56c96d9477-vps8l_2b65ecea-c14f-4702-be01-8952fabfb1be/neutron-httpd/0.log" Dec 10 20:32:31 crc kubenswrapper[4828]: I1210 20:32:31.760905 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-56c96d9477-vps8l_2b65ecea-c14f-4702-be01-8952fabfb1be/neutron-api/0.log" Dec 10 20:32:32 crc kubenswrapper[4828]: I1210 20:32:32.607871 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_5c3fe039-ce8f-4d89-8357-727c36514bf8/nova-cell0-conductor-conductor/0.log" Dec 10 20:32:32 crc kubenswrapper[4828]: I1210 20:32:32.841512 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_79e56b96-2dd9-4c50-8440-73a2bfbba5df/nova-cell1-conductor-conductor/0.log" Dec 10 20:32:32 crc kubenswrapper[4828]: I1210 20:32:32.860359 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_6e880ecd-5ad3-4da8-b817-97149eff4ec2/nova-api-log/0.log" Dec 10 20:32:33 crc kubenswrapper[4828]: I1210 20:32:33.156597 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_6e880ecd-5ad3-4da8-b817-97149eff4ec2/nova-api-api/0.log" Dec 10 20:32:33 crc kubenswrapper[4828]: I1210 20:32:33.683448 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-475xn_7dc2751a-2c10-409f-8313-f6e78fb87ba3/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:33 crc kubenswrapper[4828]: I1210 20:32:33.700994 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_78c25c19-441e-4b49-82b6-cad0ff109175/nova-cell1-novncproxy-novncproxy/0.log" Dec 10 20:32:33 crc kubenswrapper[4828]: I1210 20:32:33.968865 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b98c456a-bb60-4ef8-8367-fad750d64487/nova-metadata-log/0.log" Dec 10 20:32:34 crc kubenswrapper[4828]: I1210 20:32:34.212326 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8b0f0962-54b6-46c3-8c5d-d008874f754f/mysql-bootstrap/0.log" Dec 10 20:32:34 crc kubenswrapper[4828]: I1210 20:32:34.234400 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_1f01fc94-86ea-4b2e-b89b-7279898f9366/nova-scheduler-scheduler/0.log" Dec 10 20:32:34 crc kubenswrapper[4828]: I1210 20:32:34.420450 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8b0f0962-54b6-46c3-8c5d-d008874f754f/mysql-bootstrap/0.log" Dec 10 20:32:34 crc kubenswrapper[4828]: I1210 20:32:34.450033 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8b0f0962-54b6-46c3-8c5d-d008874f754f/galera/0.log" Dec 10 20:32:34 crc kubenswrapper[4828]: I1210 20:32:34.612314 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ddf0d133-f122-4bac-b746-222a86e29dd2/mysql-bootstrap/0.log" Dec 10 20:32:35 crc kubenswrapper[4828]: I1210 20:32:35.362992 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ddf0d133-f122-4bac-b746-222a86e29dd2/mysql-bootstrap/0.log" Dec 10 20:32:35 crc kubenswrapper[4828]: I1210 20:32:35.524180 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ddf0d133-f122-4bac-b746-222a86e29dd2/galera/0.log" Dec 10 20:32:35 crc kubenswrapper[4828]: I1210 20:32:35.803841 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_28cb9eb3-e89d-460c-9110-32e241d2a44d/openstackclient/0.log" Dec 10 20:32:35 crc kubenswrapper[4828]: I1210 20:32:35.908719 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-krphm_8fa35e66-ea0f-423d-b133-2170900e433f/ovn-controller/0.log" Dec 10 20:32:36 crc kubenswrapper[4828]: I1210 20:32:36.058426 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b98c456a-bb60-4ef8-8367-fad750d64487/nova-metadata-metadata/0.log" Dec 10 20:32:36 crc kubenswrapper[4828]: I1210 20:32:36.147066 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-zmqsz_b5204157-80d9-4397-83a5-865cf0dd293f/openstack-network-exporter/0.log" Dec 10 20:32:36 crc kubenswrapper[4828]: I1210 20:32:36.275223 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lldb4_517eff25-cf05-4abc-a46a-3d3dcd7e7f16/ovsdb-server-init/0.log" Dec 10 20:32:36 crc kubenswrapper[4828]: I1210 20:32:36.456623 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lldb4_517eff25-cf05-4abc-a46a-3d3dcd7e7f16/ovsdb-server-init/0.log" Dec 10 20:32:36 crc kubenswrapper[4828]: I1210 20:32:36.524459 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lldb4_517eff25-cf05-4abc-a46a-3d3dcd7e7f16/ovsdb-server/0.log" Dec 10 20:32:36 crc kubenswrapper[4828]: I1210 20:32:36.530087 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-lldb4_517eff25-cf05-4abc-a46a-3d3dcd7e7f16/ovs-vswitchd/0.log" Dec 10 20:32:36 crc kubenswrapper[4828]: I1210 20:32:36.699476 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-f9knx_6fb3b5f8-27e6-4d29-9b24-0fd6b0fa3f94/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:36 crc kubenswrapper[4828]: I1210 20:32:36.766647 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c323eb01-f638-4ec8-98a4-711fbd9063f2/openstack-network-exporter/0.log" Dec 10 20:32:36 crc kubenswrapper[4828]: I1210 20:32:36.820998 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c323eb01-f638-4ec8-98a4-711fbd9063f2/ovn-northd/0.log" Dec 10 20:32:36 crc kubenswrapper[4828]: I1210 20:32:36.985496 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f/ovsdbserver-nb/0.log" Dec 10 20:32:37 crc kubenswrapper[4828]: I1210 20:32:37.002261 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_1bc5f1a9-6af4-419b-a3b9-6f41b6113e8f/openstack-network-exporter/0.log" Dec 10 20:32:37 crc kubenswrapper[4828]: I1210 20:32:37.162068 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_00c4966f-831b-40d7-a49f-da33b487cd84/openstack-network-exporter/0.log" Dec 10 20:32:37 crc kubenswrapper[4828]: I1210 20:32:37.246725 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_00c4966f-831b-40d7-a49f-da33b487cd84/ovsdbserver-sb/0.log" Dec 10 20:32:37 crc kubenswrapper[4828]: I1210 20:32:37.479583 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-f568c66f6-nt8g5_06e3ba7a-1540-4093-b273-bc1c5ffa8ca6/placement-api/0.log" Dec 10 20:32:37 crc kubenswrapper[4828]: I1210 20:32:37.516264 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_5d6e68b4-1320-4111-a47d-51efcfab16b3/init-config-reloader/0.log" Dec 10 20:32:37 crc kubenswrapper[4828]: I1210 20:32:37.542084 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-f568c66f6-nt8g5_06e3ba7a-1540-4093-b273-bc1c5ffa8ca6/placement-log/0.log" Dec 10 20:32:37 crc kubenswrapper[4828]: I1210 20:32:37.710917 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_5d6e68b4-1320-4111-a47d-51efcfab16b3/init-config-reloader/0.log" Dec 10 20:32:37 crc kubenswrapper[4828]: I1210 20:32:37.729976 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_5d6e68b4-1320-4111-a47d-51efcfab16b3/config-reloader/0.log" Dec 10 20:32:37 crc kubenswrapper[4828]: I1210 20:32:37.787433 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_5d6e68b4-1320-4111-a47d-51efcfab16b3/prometheus/0.log" Dec 10 20:32:37 crc kubenswrapper[4828]: I1210 20:32:37.830345 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_5d6e68b4-1320-4111-a47d-51efcfab16b3/thanos-sidecar/0.log" Dec 10 20:32:38 crc kubenswrapper[4828]: I1210 20:32:38.029047 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3eb66f68-0443-4122-8d37-8000aefcbc3b/setup-container/0.log" Dec 10 20:32:38 crc kubenswrapper[4828]: I1210 20:32:38.223225 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3eb66f68-0443-4122-8d37-8000aefcbc3b/setup-container/0.log" Dec 10 20:32:38 crc kubenswrapper[4828]: I1210 20:32:38.253590 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3eb66f68-0443-4122-8d37-8000aefcbc3b/rabbitmq/0.log" Dec 10 20:32:38 crc kubenswrapper[4828]: I1210 20:32:38.313763 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_540f0868-0b12-44cc-806a-9d31d644da25/setup-container/0.log" Dec 10 20:32:38 crc kubenswrapper[4828]: I1210 20:32:38.561882 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_540f0868-0b12-44cc-806a-9d31d644da25/rabbitmq/0.log" Dec 10 20:32:38 crc kubenswrapper[4828]: I1210 20:32:38.569369 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_540f0868-0b12-44cc-806a-9d31d644da25/setup-container/0.log" Dec 10 20:32:38 crc kubenswrapper[4828]: I1210 20:32:38.615933 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-pjrj9_c74d14a7-f5e1-40cc-bde7-72d19e285ae9/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:38 crc kubenswrapper[4828]: I1210 20:32:38.820291 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-8psc4_12dcc39f-0cfe-498a-8de3-459c73544b77/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:38 crc kubenswrapper[4828]: I1210 20:32:38.879294 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-lcvs2_bd51b75d-403a-4b14-8f0f-72cd8d989bca/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:39 crc kubenswrapper[4828]: I1210 20:32:39.070711 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-wzcbg_d3bac884-e92c-44ec-999f-1e8ab80f4bca/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:39 crc kubenswrapper[4828]: I1210 20:32:39.387159 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-cm8mx_cba6fd50-fd63-4672-8d84-396aba02ff04/ssh-known-hosts-edpm-deployment/0.log" Dec 10 20:32:39 crc kubenswrapper[4828]: I1210 20:32:39.628846 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7b757664c-hkcjf_c6749e4d-e98e-441c-a1e8-2a1ada77eb84/proxy-server/0.log" Dec 10 20:32:39 crc kubenswrapper[4828]: I1210 20:32:39.765409 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7b757664c-hkcjf_c6749e4d-e98e-441c-a1e8-2a1ada77eb84/proxy-httpd/0.log" Dec 10 20:32:39 crc kubenswrapper[4828]: I1210 20:32:39.768998 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-bw78w_66d31859-79eb-4314-9644-455cc33ddacd/swift-ring-rebalance/0.log" Dec 10 20:32:39 crc kubenswrapper[4828]: I1210 20:32:39.971888 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e29bd910-2077-4ebe-8028-e810061fb704/account-reaper/0.log" Dec 10 20:32:39 crc kubenswrapper[4828]: I1210 20:32:39.998121 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e29bd910-2077-4ebe-8028-e810061fb704/account-auditor/0.log" Dec 10 20:32:40 crc kubenswrapper[4828]: I1210 20:32:40.043502 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e29bd910-2077-4ebe-8028-e810061fb704/account-replicator/0.log" Dec 10 20:32:40 crc kubenswrapper[4828]: I1210 20:32:40.168427 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e29bd910-2077-4ebe-8028-e810061fb704/account-server/0.log" Dec 10 20:32:40 crc kubenswrapper[4828]: I1210 20:32:40.241733 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e29bd910-2077-4ebe-8028-e810061fb704/container-auditor/0.log" Dec 10 20:32:40 crc kubenswrapper[4828]: I1210 20:32:40.283223 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e29bd910-2077-4ebe-8028-e810061fb704/container-server/0.log" Dec 10 20:32:40 crc kubenswrapper[4828]: I1210 20:32:40.284970 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e29bd910-2077-4ebe-8028-e810061fb704/container-replicator/0.log" Dec 10 20:32:40 crc kubenswrapper[4828]: I1210 20:32:40.428011 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e29bd910-2077-4ebe-8028-e810061fb704/container-updater/0.log" Dec 10 20:32:40 crc kubenswrapper[4828]: I1210 20:32:40.513567 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e29bd910-2077-4ebe-8028-e810061fb704/object-expirer/0.log" Dec 10 20:32:40 crc kubenswrapper[4828]: I1210 20:32:40.550317 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e29bd910-2077-4ebe-8028-e810061fb704/object-auditor/0.log" Dec 10 20:32:40 crc kubenswrapper[4828]: I1210 20:32:40.609546 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e29bd910-2077-4ebe-8028-e810061fb704/object-replicator/0.log" Dec 10 20:32:40 crc kubenswrapper[4828]: I1210 20:32:40.682159 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e29bd910-2077-4ebe-8028-e810061fb704/object-server/0.log" Dec 10 20:32:40 crc kubenswrapper[4828]: I1210 20:32:40.782246 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e29bd910-2077-4ebe-8028-e810061fb704/rsync/0.log" Dec 10 20:32:40 crc kubenswrapper[4828]: I1210 20:32:40.810272 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e29bd910-2077-4ebe-8028-e810061fb704/object-updater/0.log" Dec 10 20:32:40 crc kubenswrapper[4828]: I1210 20:32:40.868732 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e29bd910-2077-4ebe-8028-e810061fb704/swift-recon-cron/0.log" Dec 10 20:32:41 crc kubenswrapper[4828]: I1210 20:32:41.166186 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-csrqs_f5034a0e-652b-414a-a36f-e7f4f34da750/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:41 crc kubenswrapper[4828]: I1210 20:32:41.231263 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-power-monitoring-edpm-deployment-openstack-edpm-gcg6w_68109455-5463-42d8-9ef5-d2c667132032/telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:41 crc kubenswrapper[4828]: I1210 20:32:41.487084 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_11bbb31c-6748-4445-aa7d-bebfb5578925/test-operator-logs-container/0.log" Dec 10 20:32:41 crc kubenswrapper[4828]: I1210 20:32:41.658564 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-x8xrt_0e0b0f30-65f8-48bf-9dbf-223196457d89/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 10 20:32:42 crc kubenswrapper[4828]: I1210 20:32:42.624602 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_169c0824-0503-4c49-9980-7d045db64b8b/tempest-tests-tempest-tests-runner/0.log" Dec 10 20:32:44 crc kubenswrapper[4828]: I1210 20:32:44.788670 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:32:44 crc kubenswrapper[4828]: E1210 20:32:44.789550 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:32:46 crc kubenswrapper[4828]: I1210 20:32:46.561384 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_0c4c7fa8-3365-4ef2-abe8-e9ea79ef22ec/memcached/0.log" Dec 10 20:32:57 crc kubenswrapper[4828]: I1210 20:32:57.792651 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:32:57 crc kubenswrapper[4828]: E1210 20:32:57.793509 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:33:07 crc kubenswrapper[4828]: I1210 20:33:07.645537 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7_53cc41b8-5f56-44aa-9861-60973e96f7b7/util/0.log" Dec 10 20:33:07 crc kubenswrapper[4828]: I1210 20:33:07.893689 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7_53cc41b8-5f56-44aa-9861-60973e96f7b7/pull/0.log" Dec 10 20:33:07 crc kubenswrapper[4828]: I1210 20:33:07.894071 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7_53cc41b8-5f56-44aa-9861-60973e96f7b7/util/0.log" Dec 10 20:33:07 crc kubenswrapper[4828]: I1210 20:33:07.896013 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7_53cc41b8-5f56-44aa-9861-60973e96f7b7/pull/0.log" Dec 10 20:33:08 crc kubenswrapper[4828]: I1210 20:33:08.069659 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7_53cc41b8-5f56-44aa-9861-60973e96f7b7/util/0.log" Dec 10 20:33:08 crc kubenswrapper[4828]: I1210 20:33:08.089980 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7_53cc41b8-5f56-44aa-9861-60973e96f7b7/extract/0.log" Dec 10 20:33:08 crc kubenswrapper[4828]: I1210 20:33:08.139338 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0e5e138e65364f549de7c078512e79c32d2a2e1481f960d500560ada7fpqjw7_53cc41b8-5f56-44aa-9861-60973e96f7b7/pull/0.log" Dec 10 20:33:08 crc kubenswrapper[4828]: I1210 20:33:08.249846 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-pcfhr_b0544042-b6b7-4a9d-884a-12859ddea597/kube-rbac-proxy/0.log" Dec 10 20:33:08 crc kubenswrapper[4828]: I1210 20:33:08.350053 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-pcfhr_b0544042-b6b7-4a9d-884a-12859ddea597/manager/0.log" Dec 10 20:33:08 crc kubenswrapper[4828]: I1210 20:33:08.370155 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-cfhjs_8991af61-9866-4ae9-b7a7-7545833c4485/kube-rbac-proxy/0.log" Dec 10 20:33:08 crc kubenswrapper[4828]: I1210 20:33:08.495531 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-cfhjs_8991af61-9866-4ae9-b7a7-7545833c4485/manager/0.log" Dec 10 20:33:08 crc kubenswrapper[4828]: I1210 20:33:08.546903 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-wtcz4_bd786401-638a-4830-8a28-cc01b03d311d/kube-rbac-proxy/0.log" Dec 10 20:33:08 crc kubenswrapper[4828]: I1210 20:33:08.570287 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-wtcz4_bd786401-638a-4830-8a28-cc01b03d311d/manager/0.log" Dec 10 20:33:08 crc kubenswrapper[4828]: I1210 20:33:08.775184 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-vhf9j_1f3d156d-ddc6-47a2-9c40-5b675e535b37/kube-rbac-proxy/0.log" Dec 10 20:33:08 crc kubenswrapper[4828]: I1210 20:33:08.863386 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-vhf9j_1f3d156d-ddc6-47a2-9c40-5b675e535b37/manager/0.log" Dec 10 20:33:08 crc kubenswrapper[4828]: I1210 20:33:08.957744 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-mmqkg_bad66884-aa3d-4683-abc6-2387140d407f/kube-rbac-proxy/0.log" Dec 10 20:33:09 crc kubenswrapper[4828]: I1210 20:33:09.084128 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-mmqkg_bad66884-aa3d-4683-abc6-2387140d407f/manager/0.log" Dec 10 20:33:09 crc kubenswrapper[4828]: I1210 20:33:09.090056 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-s5jjg_e87fb7c5-676f-4549-82e3-2696306f651b/kube-rbac-proxy/0.log" Dec 10 20:33:09 crc kubenswrapper[4828]: I1210 20:33:09.205998 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-s5jjg_e87fb7c5-676f-4549-82e3-2696306f651b/manager/0.log" Dec 10 20:33:09 crc kubenswrapper[4828]: I1210 20:33:09.290250 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-x27kt_55bb5bf6-cc44-43f9-84e5-1137401fd188/kube-rbac-proxy/0.log" Dec 10 20:33:09 crc kubenswrapper[4828]: I1210 20:33:09.523036 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-f7cn8_85a6c448-c39f-4ec5-b1f7-4f7ff1bc43a0/kube-rbac-proxy/0.log" Dec 10 20:33:09 crc kubenswrapper[4828]: I1210 20:33:09.561228 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-f7cn8_85a6c448-c39f-4ec5-b1f7-4f7ff1bc43a0/manager/0.log" Dec 10 20:33:09 crc kubenswrapper[4828]: I1210 20:33:09.585765 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-x27kt_55bb5bf6-cc44-43f9-84e5-1137401fd188/manager/0.log" Dec 10 20:33:09 crc kubenswrapper[4828]: I1210 20:33:09.715261 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-4ftlx_be987ffc-f75a-4364-ac29-37fbc9bb7200/kube-rbac-proxy/0.log" Dec 10 20:33:09 crc kubenswrapper[4828]: I1210 20:33:09.804716 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-4ftlx_be987ffc-f75a-4364-ac29-37fbc9bb7200/manager/0.log" Dec 10 20:33:09 crc kubenswrapper[4828]: I1210 20:33:09.923341 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-njhhp_8f635d23-992f-4a05-9896-c38d41ec853d/kube-rbac-proxy/0.log" Dec 10 20:33:09 crc kubenswrapper[4828]: I1210 20:33:09.987819 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-njhhp_8f635d23-992f-4a05-9896-c38d41ec853d/manager/0.log" Dec 10 20:33:10 crc kubenswrapper[4828]: I1210 20:33:10.013351 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-6bqjc_23431602-592f-4c26-a5ab-1ada81fda48a/kube-rbac-proxy/0.log" Dec 10 20:33:10 crc kubenswrapper[4828]: I1210 20:33:10.186187 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-6bqjc_23431602-592f-4c26-a5ab-1ada81fda48a/manager/0.log" Dec 10 20:33:10 crc kubenswrapper[4828]: I1210 20:33:10.211019 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-smtjq_ab0dbc03-b62b-40c9-98b9-7bb192cddff3/kube-rbac-proxy/0.log" Dec 10 20:33:10 crc kubenswrapper[4828]: I1210 20:33:10.274392 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-smtjq_ab0dbc03-b62b-40c9-98b9-7bb192cddff3/manager/0.log" Dec 10 20:33:10 crc kubenswrapper[4828]: I1210 20:33:10.412471 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-j64bx_70814aec-fe59-4dc6-a085-c5a5da7b93a5/kube-rbac-proxy/0.log" Dec 10 20:33:10 crc kubenswrapper[4828]: I1210 20:33:10.528815 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-j64bx_70814aec-fe59-4dc6-a085-c5a5da7b93a5/manager/0.log" Dec 10 20:33:10 crc kubenswrapper[4828]: I1210 20:33:10.588674 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-79p7l_be915c9b-8dfe-4aab-8e74-5375e365e623/kube-rbac-proxy/0.log" Dec 10 20:33:10 crc kubenswrapper[4828]: I1210 20:33:10.711207 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-79p7l_be915c9b-8dfe-4aab-8e74-5375e365e623/manager/0.log" Dec 10 20:33:10 crc kubenswrapper[4828]: I1210 20:33:10.757444 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fk7hbh_4614b8b1-2082-4b05-8473-404f0f82772b/kube-rbac-proxy/0.log" Dec 10 20:33:10 crc kubenswrapper[4828]: I1210 20:33:10.787136 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fk7hbh_4614b8b1-2082-4b05-8473-404f0f82772b/manager/0.log" Dec 10 20:33:10 crc kubenswrapper[4828]: I1210 20:33:10.789849 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:33:10 crc kubenswrapper[4828]: E1210 20:33:10.790193 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:33:11 crc kubenswrapper[4828]: I1210 20:33:11.221212 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6b7b77566b-7lqnm_273ed82b-72f4-4fa2-b9af-e3269d82e69e/operator/0.log" Dec 10 20:33:11 crc kubenswrapper[4828]: I1210 20:33:11.513310 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-6m6c8_eaa2ffd3-6860-4f1f-9ada-04927d8a81eb/registry-server/0.log" Dec 10 20:33:11 crc kubenswrapper[4828]: I1210 20:33:11.654111 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-dxcl8_4246d7f6-cfb7-4d1d-aac0-6ba9db04a50b/kube-rbac-proxy/0.log" Dec 10 20:33:11 crc kubenswrapper[4828]: I1210 20:33:11.878759 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-88stb_3c945977-c07e-4ef4-9370-d2fa1264eee0/kube-rbac-proxy/0.log" Dec 10 20:33:11 crc kubenswrapper[4828]: I1210 20:33:11.891689 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-dxcl8_4246d7f6-cfb7-4d1d-aac0-6ba9db04a50b/manager/0.log" Dec 10 20:33:11 crc kubenswrapper[4828]: I1210 20:33:11.911897 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-88stb_3c945977-c07e-4ef4-9370-d2fa1264eee0/manager/0.log" Dec 10 20:33:12 crc kubenswrapper[4828]: I1210 20:33:12.131075 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-cv2bf_0a273a13-6393-4a24-85c8-8ac01dbf0e5c/operator/0.log" Dec 10 20:33:12 crc kubenswrapper[4828]: I1210 20:33:12.131438 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-pf44z_65429c05-8bff-4545-a085-e4c7faa29fa3/kube-rbac-proxy/0.log" Dec 10 20:33:12 crc kubenswrapper[4828]: I1210 20:33:12.339878 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-pf44z_65429c05-8bff-4545-a085-e4c7faa29fa3/manager/0.log" Dec 10 20:33:12 crc kubenswrapper[4828]: I1210 20:33:12.411947 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7b5585cf46-6j82f_25aaba8d-1069-4360-a0b4-af30543dbb02/manager/0.log" Dec 10 20:33:12 crc kubenswrapper[4828]: I1210 20:33:12.426541 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-954d6f856-2z6fx_e0825966-a2fc-4a99-9fc8-d256f885a2c5/kube-rbac-proxy/0.log" Dec 10 20:33:12 crc kubenswrapper[4828]: I1210 20:33:12.642920 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-942ph_d7200137-65c3-48b5-bc24-5cf17caa07f4/kube-rbac-proxy/0.log" Dec 10 20:33:12 crc kubenswrapper[4828]: I1210 20:33:12.695072 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-954d6f856-2z6fx_e0825966-a2fc-4a99-9fc8-d256f885a2c5/manager/0.log" Dec 10 20:33:12 crc kubenswrapper[4828]: I1210 20:33:12.701774 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-942ph_d7200137-65c3-48b5-bc24-5cf17caa07f4/manager/0.log" Dec 10 20:33:12 crc kubenswrapper[4828]: I1210 20:33:12.828460 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-s84vk_1c7fb44a-c32b-49f0-8220-9cb69be63935/manager/0.log" Dec 10 20:33:12 crc kubenswrapper[4828]: I1210 20:33:12.845046 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-75944c9b7-s84vk_1c7fb44a-c32b-49f0-8220-9cb69be63935/kube-rbac-proxy/0.log" Dec 10 20:33:22 crc kubenswrapper[4828]: I1210 20:33:22.789722 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:33:22 crc kubenswrapper[4828]: E1210 20:33:22.790487 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:33:30 crc kubenswrapper[4828]: I1210 20:33:30.509807 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-nscjn_f68aa373-3441-49f3-abc0-d5647d7bd869/control-plane-machine-set-operator/0.log" Dec 10 20:33:30 crc kubenswrapper[4828]: I1210 20:33:30.701948 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-gspnq_de5e17c4-b5df-48a1-898c-436e2c0923c3/machine-api-operator/0.log" Dec 10 20:33:30 crc kubenswrapper[4828]: I1210 20:33:30.705548 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-gspnq_de5e17c4-b5df-48a1-898c-436e2c0923c3/kube-rbac-proxy/0.log" Dec 10 20:33:35 crc kubenswrapper[4828]: I1210 20:33:35.789718 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:33:35 crc kubenswrapper[4828]: E1210 20:33:35.790752 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:33:42 crc kubenswrapper[4828]: I1210 20:33:42.645097 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-nlz6n_45d36af5-9592-4057-a3fa-00c9cd1aa9e5/cert-manager-controller/0.log" Dec 10 20:33:42 crc kubenswrapper[4828]: I1210 20:33:42.761644 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-x7cdj_eaad2b2e-c77d-4f5b-a4e7-b6fe4ae46e3a/cert-manager-cainjector/0.log" Dec 10 20:33:42 crc kubenswrapper[4828]: I1210 20:33:42.850044 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-gtg7k_6e73ff05-cad9-439a-9116-c885059594f2/cert-manager-webhook/0.log" Dec 10 20:33:47 crc kubenswrapper[4828]: I1210 20:33:47.789847 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:33:47 crc kubenswrapper[4828]: E1210 20:33:47.790839 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:33:56 crc kubenswrapper[4828]: I1210 20:33:56.377032 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-bdm9r_6c267cf9-bc24-4e82-8293-464d90e3dbc2/nmstate-console-plugin/0.log" Dec 10 20:33:56 crc kubenswrapper[4828]: I1210 20:33:56.578725 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-5tn4b_ce59be34-eb49-4a48-b2cb-e627bf2ed853/nmstate-handler/0.log" Dec 10 20:33:56 crc kubenswrapper[4828]: I1210 20:33:56.603672 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-rz5wn_9f6baf7f-f16b-4acf-a2c5-18d72615ff42/nmstate-metrics/0.log" Dec 10 20:33:56 crc kubenswrapper[4828]: I1210 20:33:56.643740 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-rz5wn_9f6baf7f-f16b-4acf-a2c5-18d72615ff42/kube-rbac-proxy/0.log" Dec 10 20:33:56 crc kubenswrapper[4828]: I1210 20:33:56.787182 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-d4x88_9c4e3c85-4f86-434a-8adb-d22d17b3b44c/nmstate-operator/0.log" Dec 10 20:33:56 crc kubenswrapper[4828]: I1210 20:33:56.849390 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-j9tv4_afc9bf44-dee0-4c71-8e9a-519f382d1857/nmstate-webhook/0.log" Dec 10 20:33:59 crc kubenswrapper[4828]: I1210 20:33:59.803474 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:33:59 crc kubenswrapper[4828]: E1210 20:33:59.804595 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:34:10 crc kubenswrapper[4828]: I1210 20:34:10.057072 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-9c7bf5466-2tc42_f4ae06be-8676-4704-aa09-b27912dd5dbf/kube-rbac-proxy/0.log" Dec 10 20:34:10 crc kubenswrapper[4828]: I1210 20:34:10.210616 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-9c7bf5466-2tc42_f4ae06be-8676-4704-aa09-b27912dd5dbf/manager/0.log" Dec 10 20:34:11 crc kubenswrapper[4828]: I1210 20:34:11.789385 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:34:11 crc kubenswrapper[4828]: E1210 20:34:11.790274 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:34:23 crc kubenswrapper[4828]: I1210 20:34:23.789926 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:34:23 crc kubenswrapper[4828]: E1210 20:34:23.790697 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:34:23 crc kubenswrapper[4828]: I1210 20:34:23.826182 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-b94hr_fe41e1e4-96fb-4798-9dc8-3382d8c74b46/cluster-logging-operator/0.log" Dec 10 20:34:23 crc kubenswrapper[4828]: I1210 20:34:23.967006 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-bwvf2_416a301b-b484-4f47-9901-6c033705c501/collector/0.log" Dec 10 20:34:24 crc kubenswrapper[4828]: I1210 20:34:24.021536 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_d3501ce8-e4b7-420d-b727-5ed3015f39d2/loki-compactor/0.log" Dec 10 20:34:24 crc kubenswrapper[4828]: I1210 20:34:24.169330 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-lg62n_14db470d-4202-4b2d-9339-65ba24e521d0/loki-distributor/0.log" Dec 10 20:34:24 crc kubenswrapper[4828]: I1210 20:34:24.211310 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-6c4d974455-7lpjj_cf17217c-0dc6-4edd-8334-8896f103f9f1/gateway/0.log" Dec 10 20:34:24 crc kubenswrapper[4828]: I1210 20:34:24.247317 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-6c4d974455-7lpjj_cf17217c-0dc6-4edd-8334-8896f103f9f1/opa/0.log" Dec 10 20:34:24 crc kubenswrapper[4828]: I1210 20:34:24.391073 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-6c4d974455-7r2h2_4c744ad4-db8c-42ad-9b49-ecb32b023cb6/gateway/0.log" Dec 10 20:34:24 crc kubenswrapper[4828]: I1210 20:34:24.400189 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-6c4d974455-7r2h2_4c744ad4-db8c-42ad-9b49-ecb32b023cb6/opa/0.log" Dec 10 20:34:24 crc kubenswrapper[4828]: I1210 20:34:24.534584 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_df650a21-47af-445a-9b8a-8908684aa6b9/loki-index-gateway/0.log" Dec 10 20:34:24 crc kubenswrapper[4828]: I1210 20:34:24.680334 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_a4483694-bd44-4582-80b0-17c774d39682/loki-ingester/0.log" Dec 10 20:34:24 crc kubenswrapper[4828]: I1210 20:34:24.729189 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-4pkdl_15695070-e7cd-4af1-980d-e79defa1efaf/loki-querier/0.log" Dec 10 20:34:24 crc kubenswrapper[4828]: I1210 20:34:24.822534 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-qqdqb_6cf6784a-0699-4d1c-ba27-cb8f25ab729b/loki-query-frontend/0.log" Dec 10 20:34:38 crc kubenswrapper[4828]: I1210 20:34:38.268747 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-h4cp8_0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c/kube-rbac-proxy/0.log" Dec 10 20:34:38 crc kubenswrapper[4828]: I1210 20:34:38.444562 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-h4cp8_0aeb78a5-f8f6-4cdc-ba8f-32d5fc763d6c/controller/0.log" Dec 10 20:34:38 crc kubenswrapper[4828]: I1210 20:34:38.479755 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/cp-frr-files/0.log" Dec 10 20:34:38 crc kubenswrapper[4828]: I1210 20:34:38.581191 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/cp-frr-files/0.log" Dec 10 20:34:38 crc kubenswrapper[4828]: I1210 20:34:38.612070 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/cp-reloader/0.log" Dec 10 20:34:38 crc kubenswrapper[4828]: I1210 20:34:38.620788 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/cp-metrics/0.log" Dec 10 20:34:38 crc kubenswrapper[4828]: I1210 20:34:38.677155 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/cp-reloader/0.log" Dec 10 20:34:38 crc kubenswrapper[4828]: I1210 20:34:38.789624 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:34:38 crc kubenswrapper[4828]: E1210 20:34:38.789901 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:34:38 crc kubenswrapper[4828]: I1210 20:34:38.851860 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/cp-reloader/0.log" Dec 10 20:34:38 crc kubenswrapper[4828]: I1210 20:34:38.870193 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/cp-frr-files/0.log" Dec 10 20:34:38 crc kubenswrapper[4828]: I1210 20:34:38.875821 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/cp-metrics/0.log" Dec 10 20:34:38 crc kubenswrapper[4828]: I1210 20:34:38.895082 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/cp-metrics/0.log" Dec 10 20:34:39 crc kubenswrapper[4828]: I1210 20:34:39.027260 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/cp-frr-files/0.log" Dec 10 20:34:39 crc kubenswrapper[4828]: I1210 20:34:39.059192 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/cp-reloader/0.log" Dec 10 20:34:39 crc kubenswrapper[4828]: I1210 20:34:39.072471 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/cp-metrics/0.log" Dec 10 20:34:39 crc kubenswrapper[4828]: I1210 20:34:39.079781 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/controller/0.log" Dec 10 20:34:39 crc kubenswrapper[4828]: I1210 20:34:39.263755 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/frr-metrics/0.log" Dec 10 20:34:39 crc kubenswrapper[4828]: I1210 20:34:39.271634 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/kube-rbac-proxy/0.log" Dec 10 20:34:39 crc kubenswrapper[4828]: I1210 20:34:39.282445 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/kube-rbac-proxy-frr/0.log" Dec 10 20:34:39 crc kubenswrapper[4828]: I1210 20:34:39.523193 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/reloader/0.log" Dec 10 20:34:39 crc kubenswrapper[4828]: I1210 20:34:39.537744 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-hzjlj_159d6965-2b9c-461f-8f20-cf338919fcd7/frr-k8s-webhook-server/0.log" Dec 10 20:34:39 crc kubenswrapper[4828]: I1210 20:34:39.749398 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-75b57b9fc5-tdblw_4a495b9b-2ebd-43bd-bb0c-64f32ad257ae/manager/0.log" Dec 10 20:34:39 crc kubenswrapper[4828]: I1210 20:34:39.985299 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5bd8d89b5d-tdlk8_87eba563-d95b-486f-bde7-61d7a764d996/webhook-server/0.log" Dec 10 20:34:40 crc kubenswrapper[4828]: I1210 20:34:40.031211 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-wjndz_dd16c5ff-d9b5-4df0-a3ca-b7900966bc37/kube-rbac-proxy/0.log" Dec 10 20:34:40 crc kubenswrapper[4828]: I1210 20:34:40.738009 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-wjndz_dd16c5ff-d9b5-4df0-a3ca-b7900966bc37/speaker/0.log" Dec 10 20:34:41 crc kubenswrapper[4828]: I1210 20:34:41.182647 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-97pxg_74aa2e85-8265-4d0b-bd8e-3d1c13c7dfab/frr/0.log" Dec 10 20:34:52 crc kubenswrapper[4828]: I1210 20:34:52.789780 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:34:53 crc kubenswrapper[4828]: I1210 20:34:53.908841 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"325ae6fd60148173e9b100cf5cb87a1e61d379c9db65348aeb2ae47852a05290"} Dec 10 20:34:54 crc kubenswrapper[4828]: I1210 20:34:54.377035 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws_2c2bf8e7-87cf-4ede-bc49-e955290a66ee/util/0.log" Dec 10 20:34:54 crc kubenswrapper[4828]: I1210 20:34:54.633128 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws_2c2bf8e7-87cf-4ede-bc49-e955290a66ee/util/0.log" Dec 10 20:34:54 crc kubenswrapper[4828]: I1210 20:34:54.661451 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws_2c2bf8e7-87cf-4ede-bc49-e955290a66ee/pull/0.log" Dec 10 20:34:54 crc kubenswrapper[4828]: I1210 20:34:54.735657 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws_2c2bf8e7-87cf-4ede-bc49-e955290a66ee/pull/0.log" Dec 10 20:34:54 crc kubenswrapper[4828]: I1210 20:34:54.861170 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws_2c2bf8e7-87cf-4ede-bc49-e955290a66ee/pull/0.log" Dec 10 20:34:54 crc kubenswrapper[4828]: I1210 20:34:54.868076 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws_2c2bf8e7-87cf-4ede-bc49-e955290a66ee/extract/0.log" Dec 10 20:34:54 crc kubenswrapper[4828]: I1210 20:34:54.894005 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8js9ws_2c2bf8e7-87cf-4ede-bc49-e955290a66ee/util/0.log" Dec 10 20:34:55 crc kubenswrapper[4828]: I1210 20:34:55.041968 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd_7d03ca8f-a473-43a1-a97e-9b141ab54ee3/util/0.log" Dec 10 20:34:55 crc kubenswrapper[4828]: I1210 20:34:55.323360 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd_7d03ca8f-a473-43a1-a97e-9b141ab54ee3/util/0.log" Dec 10 20:34:55 crc kubenswrapper[4828]: I1210 20:34:55.388228 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd_7d03ca8f-a473-43a1-a97e-9b141ab54ee3/pull/0.log" Dec 10 20:34:55 crc kubenswrapper[4828]: I1210 20:34:55.396024 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd_7d03ca8f-a473-43a1-a97e-9b141ab54ee3/pull/0.log" Dec 10 20:34:55 crc kubenswrapper[4828]: I1210 20:34:55.561688 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd_7d03ca8f-a473-43a1-a97e-9b141ab54ee3/util/0.log" Dec 10 20:34:55 crc kubenswrapper[4828]: I1210 20:34:55.594368 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd_7d03ca8f-a473-43a1-a97e-9b141ab54ee3/pull/0.log" Dec 10 20:34:55 crc kubenswrapper[4828]: I1210 20:34:55.623901 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fn9sbd_7d03ca8f-a473-43a1-a97e-9b141ab54ee3/extract/0.log" Dec 10 20:34:55 crc kubenswrapper[4828]: I1210 20:34:55.825624 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f_025427dc-d729-4b53-b665-de357c409d97/util/0.log" Dec 10 20:34:55 crc kubenswrapper[4828]: I1210 20:34:55.987406 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f_025427dc-d729-4b53-b665-de357c409d97/util/0.log" Dec 10 20:34:56 crc kubenswrapper[4828]: I1210 20:34:56.042968 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f_025427dc-d729-4b53-b665-de357c409d97/pull/0.log" Dec 10 20:34:56 crc kubenswrapper[4828]: I1210 20:34:56.051548 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f_025427dc-d729-4b53-b665-de357c409d97/pull/0.log" Dec 10 20:34:56 crc kubenswrapper[4828]: I1210 20:34:56.247317 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f_025427dc-d729-4b53-b665-de357c409d97/util/0.log" Dec 10 20:34:56 crc kubenswrapper[4828]: I1210 20:34:56.664190 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f_025427dc-d729-4b53-b665-de357c409d97/extract/0.log" Dec 10 20:34:56 crc kubenswrapper[4828]: I1210 20:34:56.704607 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210zzv6f_025427dc-d729-4b53-b665-de357c409d97/pull/0.log" Dec 10 20:34:56 crc kubenswrapper[4828]: I1210 20:34:56.886406 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5_ef8a4983-2aae-4e7e-a416-317ffceb7b53/util/0.log" Dec 10 20:34:57 crc kubenswrapper[4828]: I1210 20:34:57.086085 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5_ef8a4983-2aae-4e7e-a416-317ffceb7b53/pull/0.log" Dec 10 20:34:57 crc kubenswrapper[4828]: I1210 20:34:57.109997 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5_ef8a4983-2aae-4e7e-a416-317ffceb7b53/util/0.log" Dec 10 20:34:57 crc kubenswrapper[4828]: I1210 20:34:57.170093 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5_ef8a4983-2aae-4e7e-a416-317ffceb7b53/pull/0.log" Dec 10 20:34:57 crc kubenswrapper[4828]: I1210 20:34:57.360508 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5_ef8a4983-2aae-4e7e-a416-317ffceb7b53/util/0.log" Dec 10 20:34:57 crc kubenswrapper[4828]: I1210 20:34:57.399888 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5_ef8a4983-2aae-4e7e-a416-317ffceb7b53/pull/0.log" Dec 10 20:34:57 crc kubenswrapper[4828]: I1210 20:34:57.400492 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fz9lb5_ef8a4983-2aae-4e7e-a416-317ffceb7b53/extract/0.log" Dec 10 20:34:57 crc kubenswrapper[4828]: I1210 20:34:57.585982 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9_1f7d42e8-5aae-40e3-a5f8-bd398b69db43/util/0.log" Dec 10 20:34:57 crc kubenswrapper[4828]: I1210 20:34:57.862624 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9_1f7d42e8-5aae-40e3-a5f8-bd398b69db43/util/0.log" Dec 10 20:34:57 crc kubenswrapper[4828]: I1210 20:34:57.917422 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9_1f7d42e8-5aae-40e3-a5f8-bd398b69db43/pull/0.log" Dec 10 20:34:57 crc kubenswrapper[4828]: I1210 20:34:57.919566 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9_1f7d42e8-5aae-40e3-a5f8-bd398b69db43/pull/0.log" Dec 10 20:34:58 crc kubenswrapper[4828]: I1210 20:34:58.143629 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9_1f7d42e8-5aae-40e3-a5f8-bd398b69db43/pull/0.log" Dec 10 20:34:58 crc kubenswrapper[4828]: I1210 20:34:58.149171 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9_1f7d42e8-5aae-40e3-a5f8-bd398b69db43/util/0.log" Dec 10 20:34:58 crc kubenswrapper[4828]: I1210 20:34:58.187087 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lclh9_1f7d42e8-5aae-40e3-a5f8-bd398b69db43/extract/0.log" Dec 10 20:34:58 crc kubenswrapper[4828]: I1210 20:34:58.343861 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-df2d5_fb55e51a-e3e7-4bea-8791-aec28f2b8873/extract-utilities/0.log" Dec 10 20:34:58 crc kubenswrapper[4828]: I1210 20:34:58.539752 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-df2d5_fb55e51a-e3e7-4bea-8791-aec28f2b8873/extract-utilities/0.log" Dec 10 20:34:58 crc kubenswrapper[4828]: I1210 20:34:58.541833 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-df2d5_fb55e51a-e3e7-4bea-8791-aec28f2b8873/extract-content/0.log" Dec 10 20:34:58 crc kubenswrapper[4828]: I1210 20:34:58.544364 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-df2d5_fb55e51a-e3e7-4bea-8791-aec28f2b8873/extract-content/0.log" Dec 10 20:34:58 crc kubenswrapper[4828]: I1210 20:34:58.696589 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-df2d5_fb55e51a-e3e7-4bea-8791-aec28f2b8873/extract-utilities/0.log" Dec 10 20:34:58 crc kubenswrapper[4828]: I1210 20:34:58.745469 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-df2d5_fb55e51a-e3e7-4bea-8791-aec28f2b8873/extract-content/0.log" Dec 10 20:34:58 crc kubenswrapper[4828]: I1210 20:34:58.799993 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t9xj4_ca31d980-917d-4956-adb9-e31050c30f70/extract-utilities/0.log" Dec 10 20:34:59 crc kubenswrapper[4828]: I1210 20:34:59.054385 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t9xj4_ca31d980-917d-4956-adb9-e31050c30f70/extract-content/0.log" Dec 10 20:34:59 crc kubenswrapper[4828]: I1210 20:34:59.223965 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t9xj4_ca31d980-917d-4956-adb9-e31050c30f70/extract-utilities/0.log" Dec 10 20:34:59 crc kubenswrapper[4828]: I1210 20:34:59.282996 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t9xj4_ca31d980-917d-4956-adb9-e31050c30f70/extract-content/0.log" Dec 10 20:34:59 crc kubenswrapper[4828]: I1210 20:34:59.386777 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-df2d5_fb55e51a-e3e7-4bea-8791-aec28f2b8873/registry-server/0.log" Dec 10 20:34:59 crc kubenswrapper[4828]: I1210 20:34:59.424758 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t9xj4_ca31d980-917d-4956-adb9-e31050c30f70/extract-content/0.log" Dec 10 20:34:59 crc kubenswrapper[4828]: I1210 20:34:59.435854 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t9xj4_ca31d980-917d-4956-adb9-e31050c30f70/extract-utilities/0.log" Dec 10 20:34:59 crc kubenswrapper[4828]: I1210 20:34:59.647621 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-f74sl_8df288f1-1cd4-4381-b519-5a353cd63c59/marketplace-operator/0.log" Dec 10 20:34:59 crc kubenswrapper[4828]: I1210 20:34:59.696629 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hm6hf_b49ed428-86d9-4b1d-a2c6-f3f79fadb13d/extract-utilities/0.log" Dec 10 20:34:59 crc kubenswrapper[4828]: I1210 20:34:59.925872 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hm6hf_b49ed428-86d9-4b1d-a2c6-f3f79fadb13d/extract-utilities/0.log" Dec 10 20:34:59 crc kubenswrapper[4828]: I1210 20:34:59.946738 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hm6hf_b49ed428-86d9-4b1d-a2c6-f3f79fadb13d/extract-content/0.log" Dec 10 20:34:59 crc kubenswrapper[4828]: I1210 20:34:59.973266 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hm6hf_b49ed428-86d9-4b1d-a2c6-f3f79fadb13d/extract-content/0.log" Dec 10 20:35:00 crc kubenswrapper[4828]: I1210 20:35:00.208967 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hm6hf_b49ed428-86d9-4b1d-a2c6-f3f79fadb13d/extract-content/0.log" Dec 10 20:35:00 crc kubenswrapper[4828]: I1210 20:35:00.260217 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-t9xj4_ca31d980-917d-4956-adb9-e31050c30f70/registry-server/0.log" Dec 10 20:35:00 crc kubenswrapper[4828]: I1210 20:35:00.469483 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hm6hf_b49ed428-86d9-4b1d-a2c6-f3f79fadb13d/extract-utilities/0.log" Dec 10 20:35:00 crc kubenswrapper[4828]: I1210 20:35:00.678724 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hm6hf_b49ed428-86d9-4b1d-a2c6-f3f79fadb13d/registry-server/0.log" Dec 10 20:35:00 crc kubenswrapper[4828]: I1210 20:35:00.707782 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9jxhs_ead62f92-39c7-46bb-be5d-f721322ee339/extract-utilities/0.log" Dec 10 20:35:00 crc kubenswrapper[4828]: I1210 20:35:00.723360 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9jxhs_ead62f92-39c7-46bb-be5d-f721322ee339/extract-utilities/0.log" Dec 10 20:35:00 crc kubenswrapper[4828]: I1210 20:35:00.784477 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9jxhs_ead62f92-39c7-46bb-be5d-f721322ee339/extract-content/0.log" Dec 10 20:35:00 crc kubenswrapper[4828]: I1210 20:35:00.794489 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9jxhs_ead62f92-39c7-46bb-be5d-f721322ee339/extract-content/0.log" Dec 10 20:35:00 crc kubenswrapper[4828]: I1210 20:35:00.934088 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9jxhs_ead62f92-39c7-46bb-be5d-f721322ee339/extract-utilities/0.log" Dec 10 20:35:00 crc kubenswrapper[4828]: I1210 20:35:00.971902 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9jxhs_ead62f92-39c7-46bb-be5d-f721322ee339/extract-content/0.log" Dec 10 20:35:01 crc kubenswrapper[4828]: I1210 20:35:01.309309 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-9jxhs_ead62f92-39c7-46bb-be5d-f721322ee339/registry-server/0.log" Dec 10 20:35:13 crc kubenswrapper[4828]: I1210 20:35:13.970482 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-vpjkk_fe95ea54-da09-4ec9-a930-5be068b09954/prometheus-operator/0.log" Dec 10 20:35:14 crc kubenswrapper[4828]: I1210 20:35:14.156933 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6774fff8-2glrf_68b3a59c-1ec9-4381-a92a-810b554d76e0/prometheus-operator-admission-webhook/0.log" Dec 10 20:35:14 crc kubenswrapper[4828]: I1210 20:35:14.237851 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-6774fff8-sbnvz_b5194787-06b3-40a3-92dd-7a7471ba5829/prometheus-operator-admission-webhook/0.log" Dec 10 20:35:14 crc kubenswrapper[4828]: I1210 20:35:14.354881 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-flqjw_b55b0a41-ae3f-4acf-bf2d-0378868dca03/operator/0.log" Dec 10 20:35:14 crc kubenswrapper[4828]: I1210 20:35:14.443944 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-7d5fb4cbfb-nq29p_6624e745-c1ca-4dfd-9ee8-107310f3d368/observability-ui-dashboards/0.log" Dec 10 20:35:14 crc kubenswrapper[4828]: I1210 20:35:14.560487 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-57l2h_ebb4036a-1bd8-4636-8bea-63b0a6b4cc38/perses-operator/0.log" Dec 10 20:35:27 crc kubenswrapper[4828]: I1210 20:35:27.824331 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-9c7bf5466-2tc42_f4ae06be-8676-4704-aa09-b27912dd5dbf/kube-rbac-proxy/0.log" Dec 10 20:35:27 crc kubenswrapper[4828]: I1210 20:35:27.856113 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-9c7bf5466-2tc42_f4ae06be-8676-4704-aa09-b27912dd5dbf/manager/0.log" Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.521670 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="5d6e68b4-1320-4111-a47d-51efcfab16b3" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.161:9090/-/healthy\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.521870 4828 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.009596208s: [/var/lib/containers/storage/overlay/5a538bbed1de1bf56a0f5cc6179c7dea08b077a6ca8c88a472aac741af8922bf/diff /var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-mmqkg_bad66884-aa3d-4683-abc6-2387140d407f/kube-rbac-proxy/0.log]; will not log again for this container unless duration exceeds 2s Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.521986 4828 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 4.002739016s: [/var/lib/containers/storage/overlay/9e212dd5f99bb7c7a55e3dedf443616dacd6457c4876dbe0521293b06bc19305/diff /var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-njhhp_8f635d23-992f-4a05-9896-c38d41ec853d/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.522010 4828 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.026524431s: [/var/lib/containers/storage/overlay/eb7f98dc6d8658bcade876ea063c013fbb51601e2ca10a489c4b7228927cb69d/diff /var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-f7cn8_85a6c448-c39f-4ec5-b1f7-4f7ff1bc43a0/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.522001 4828 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.011082688s: [/var/lib/containers/storage/overlay/c6b98da775300841ee9ceeb5c07caa879eee657b66032c02a689917024514c0a/diff /var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-4pkdl_15695070-e7cd-4af1-980d-e79defa1efaf/loki-querier/0.log]; will not log again for this container unless duration exceeds 2s Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.522066 4828 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.023059308s: [/var/lib/containers/storage/overlay/0e7cadc72b72c0a4b5e9b7c1645fdd152caa6c431dade2ab156eaf0c59880f8d/diff /var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-qqdqb_6cf6784a-0699-4d1c-ba27-cb8f25ab729b/loki-query-frontend/0.log]; will not log again for this container unless duration exceeds 2s Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.522068 4828 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 4.007076382s: [/var/lib/containers/storage/overlay/899c7809568439dbb8a6f22039cb351b5568973a665b44948ca6d6edeadd645c/diff /var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-mmqkg_bad66884-aa3d-4683-abc6-2387140d407f/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.522244 4828 trace.go:236] Trace[1945727526]: "Calculate volume metrics of env-overrides for pod openshift-network-node-identity/network-node-identity-vrzqb" (10-Dec-2025 20:36:06.311) (total time: 3206ms): Dec 10 20:36:09 crc kubenswrapper[4828]: Trace[1945727526]: [3.206620536s] [3.206620536s] END Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.522697 4828 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-wjndz" podUID="dd16c5ff-d9b5-4df0-a3ca-b7900966bc37" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.522368 4828 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.017115309s: [/var/lib/containers/storage/overlay/4bad498533ce6904b06d373142424d3698ca203bf6fc13b68c7bd8dddd462fba/diff /var/log/pods/openshift-logging_logging-loki-index-gateway-0_df650a21-47af-445a-9b8a-8908684aa6b9/loki-index-gateway/0.log]; will not log again for this container unless duration exceeds 2s Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.523663 4828 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 4.004171925s: [/var/lib/containers/storage/overlay/c3d241cbc0c7446400263b1db286ff13dac6b955a01b574d2433f528e109fbd7/diff /var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-6bqjc_23431602-592f-4c26-a5ab-1ada81fda48a/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.523716 4828 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.012815194s: [/var/lib/containers/storage/overlay/46144d80115cfd3122463d5c425dc723218fa332377d672f0c0c1209954c9e99/diff /var/log/pods/openshift-logging_logging-loki-gateway-6c4d974455-7lpjj_cf17217c-0dc6-4edd-8334-8896f103f9f1/gateway/0.log]; will not log again for this container unless duration exceeds 2s Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.523778 4828 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.043120925s: [/var/lib/containers/storage/overlay/6c3d6221dcd7f6b3d9505c4a6cc16cdf936f94f045b0795c63085b4e37268468/diff /var/log/pods/openshift-logging_logging-loki-ingester-0_a4483694-bd44-4582-80b0-17c774d39682/loki-ingester/0.log]; will not log again for this container unless duration exceeds 2s Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.523851 4828 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 4.00435517s: [/var/lib/containers/storage/overlay/7cbf248c6df10c29c4428d06e8e90f2fbec8950c0dbf61e5c3f5357ac271a038/diff /var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-wtcz4_bd786401-638a-4830-8a28-cc01b03d311d/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.524046 4828 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.023406077s: [/var/lib/containers/storage/overlay/78fac74dcf499e97379c617dd9540acabfe196f7871ad4738008c49613087dc7/diff /var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-pcfhr_b0544042-b6b7-4a9d-884a-12859ddea597/kube-rbac-proxy/0.log]; will not log again for this container unless duration exceeds 2s Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.524511 4828 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 4.426656178s: [/var/lib/containers/storage/overlay/a76893040bb9c9e90ec53fad2542ae5188a38d0648f7091855b4217964f9d306/diff /var/log/pods/openstack_heat-cfnapi-cb5d96886-n5cp9_d1887db4-171d-4bd1-8337-4a8c72e9e93d/heat-cfnapi/0.log]; will not log again for this container unless duration exceeds 2s Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.524578 4828 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.024788785s: [/var/lib/containers/storage/overlay/ee741a23c55a0a8b292cda42c3fe04b8b035cc67ef25d774d442b316ca1e9aad/diff /var/log/pods/openshift-logging_logging-loki-gateway-6c4d974455-7r2h2_4c744ad4-db8c-42ad-9b49-ecb32b023cb6/gateway/0.log]; will not log again for this container unless duration exceeds 2s Dec 10 20:36:09 crc kubenswrapper[4828]: I1210 20:36:09.524920 4828 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 4.426998767s: [/var/lib/containers/storage/overlay/3dae5aec266e7b9d6f5c44b30d1e2cb799b4308fe96e03f1344602a4a03921ac/diff /var/log/pods/openstack_heat-api-5859bd5774-r9zl7_961e506a-98a1-48e1-afd5-c4326af0bdb3/heat-api/0.log]; will not log again for this container unless duration exceeds 2s Dec 10 20:37:00 crc kubenswrapper[4828]: I1210 20:37:00.011562 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fq8zv"] Dec 10 20:37:00 crc kubenswrapper[4828]: E1210 20:37:00.012966 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54ac3720-5d29-4c80-885a-33ce3a73bdf2" containerName="container-00" Dec 10 20:37:00 crc kubenswrapper[4828]: I1210 20:37:00.012986 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="54ac3720-5d29-4c80-885a-33ce3a73bdf2" containerName="container-00" Dec 10 20:37:00 crc kubenswrapper[4828]: I1210 20:37:00.013294 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="54ac3720-5d29-4c80-885a-33ce3a73bdf2" containerName="container-00" Dec 10 20:37:00 crc kubenswrapper[4828]: I1210 20:37:00.015941 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fq8zv" Dec 10 20:37:00 crc kubenswrapper[4828]: I1210 20:37:00.035203 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fq8zv"] Dec 10 20:37:00 crc kubenswrapper[4828]: I1210 20:37:00.173517 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xntbw\" (UniqueName: \"kubernetes.io/projected/3fefba56-6955-484a-a03f-7d46ffe45540-kube-api-access-xntbw\") pod \"certified-operators-fq8zv\" (UID: \"3fefba56-6955-484a-a03f-7d46ffe45540\") " pod="openshift-marketplace/certified-operators-fq8zv" Dec 10 20:37:00 crc kubenswrapper[4828]: I1210 20:37:00.173933 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fefba56-6955-484a-a03f-7d46ffe45540-utilities\") pod \"certified-operators-fq8zv\" (UID: \"3fefba56-6955-484a-a03f-7d46ffe45540\") " pod="openshift-marketplace/certified-operators-fq8zv" Dec 10 20:37:00 crc kubenswrapper[4828]: I1210 20:37:00.174177 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fefba56-6955-484a-a03f-7d46ffe45540-catalog-content\") pod \"certified-operators-fq8zv\" (UID: \"3fefba56-6955-484a-a03f-7d46ffe45540\") " pod="openshift-marketplace/certified-operators-fq8zv" Dec 10 20:37:00 crc kubenswrapper[4828]: I1210 20:37:00.276152 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fefba56-6955-484a-a03f-7d46ffe45540-utilities\") pod \"certified-operators-fq8zv\" (UID: \"3fefba56-6955-484a-a03f-7d46ffe45540\") " pod="openshift-marketplace/certified-operators-fq8zv" Dec 10 20:37:00 crc kubenswrapper[4828]: I1210 20:37:00.276307 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fefba56-6955-484a-a03f-7d46ffe45540-catalog-content\") pod \"certified-operators-fq8zv\" (UID: \"3fefba56-6955-484a-a03f-7d46ffe45540\") " pod="openshift-marketplace/certified-operators-fq8zv" Dec 10 20:37:00 crc kubenswrapper[4828]: I1210 20:37:00.276355 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xntbw\" (UniqueName: \"kubernetes.io/projected/3fefba56-6955-484a-a03f-7d46ffe45540-kube-api-access-xntbw\") pod \"certified-operators-fq8zv\" (UID: \"3fefba56-6955-484a-a03f-7d46ffe45540\") " pod="openshift-marketplace/certified-operators-fq8zv" Dec 10 20:37:00 crc kubenswrapper[4828]: I1210 20:37:00.277060 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fefba56-6955-484a-a03f-7d46ffe45540-catalog-content\") pod \"certified-operators-fq8zv\" (UID: \"3fefba56-6955-484a-a03f-7d46ffe45540\") " pod="openshift-marketplace/certified-operators-fq8zv" Dec 10 20:37:00 crc kubenswrapper[4828]: I1210 20:37:00.277443 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fefba56-6955-484a-a03f-7d46ffe45540-utilities\") pod \"certified-operators-fq8zv\" (UID: \"3fefba56-6955-484a-a03f-7d46ffe45540\") " pod="openshift-marketplace/certified-operators-fq8zv" Dec 10 20:37:00 crc kubenswrapper[4828]: I1210 20:37:00.298460 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xntbw\" (UniqueName: \"kubernetes.io/projected/3fefba56-6955-484a-a03f-7d46ffe45540-kube-api-access-xntbw\") pod \"certified-operators-fq8zv\" (UID: \"3fefba56-6955-484a-a03f-7d46ffe45540\") " pod="openshift-marketplace/certified-operators-fq8zv" Dec 10 20:37:00 crc kubenswrapper[4828]: I1210 20:37:00.355884 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fq8zv" Dec 10 20:37:00 crc kubenswrapper[4828]: I1210 20:37:00.969979 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fq8zv"] Dec 10 20:37:01 crc kubenswrapper[4828]: I1210 20:37:01.281618 4828 generic.go:334] "Generic (PLEG): container finished" podID="3fefba56-6955-484a-a03f-7d46ffe45540" containerID="d1ed28c7d79f3b3b62f3d6133061e9417bfdc585a19fe277dedfb95cdd1e7fa6" exitCode=0 Dec 10 20:37:01 crc kubenswrapper[4828]: I1210 20:37:01.281733 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fq8zv" event={"ID":"3fefba56-6955-484a-a03f-7d46ffe45540","Type":"ContainerDied","Data":"d1ed28c7d79f3b3b62f3d6133061e9417bfdc585a19fe277dedfb95cdd1e7fa6"} Dec 10 20:37:01 crc kubenswrapper[4828]: I1210 20:37:01.283055 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fq8zv" event={"ID":"3fefba56-6955-484a-a03f-7d46ffe45540","Type":"ContainerStarted","Data":"ab15b656ac21c1c876f769b2fb34f30972058560e91387ec0fe5f491abef419b"} Dec 10 20:37:01 crc kubenswrapper[4828]: I1210 20:37:01.284625 4828 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 10 20:37:02 crc kubenswrapper[4828]: I1210 20:37:02.300046 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fq8zv" event={"ID":"3fefba56-6955-484a-a03f-7d46ffe45540","Type":"ContainerStarted","Data":"28517f630bc0a5ec40ea949b3a2bac46e3d9b4de89386cd457579ff31cee70fb"} Dec 10 20:37:03 crc kubenswrapper[4828]: I1210 20:37:03.313660 4828 generic.go:334] "Generic (PLEG): container finished" podID="3fefba56-6955-484a-a03f-7d46ffe45540" containerID="28517f630bc0a5ec40ea949b3a2bac46e3d9b4de89386cd457579ff31cee70fb" exitCode=0 Dec 10 20:37:03 crc kubenswrapper[4828]: I1210 20:37:03.313712 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fq8zv" event={"ID":"3fefba56-6955-484a-a03f-7d46ffe45540","Type":"ContainerDied","Data":"28517f630bc0a5ec40ea949b3a2bac46e3d9b4de89386cd457579ff31cee70fb"} Dec 10 20:37:04 crc kubenswrapper[4828]: I1210 20:37:04.329127 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fq8zv" event={"ID":"3fefba56-6955-484a-a03f-7d46ffe45540","Type":"ContainerStarted","Data":"33054382e319c3e206537433bc39aa8e9f39d8ac4571f9f445f5e1baf7bc46b0"} Dec 10 20:37:04 crc kubenswrapper[4828]: I1210 20:37:04.365260 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fq8zv" podStartSLOduration=2.918560879 podStartE2EDuration="5.365239541s" podCreationTimestamp="2025-12-10 20:36:59 +0000 UTC" firstStartedPulling="2025-12-10 20:37:01.283328513 +0000 UTC m=+6101.793939518" lastFinishedPulling="2025-12-10 20:37:03.730007175 +0000 UTC m=+6104.240618180" observedRunningTime="2025-12-10 20:37:04.353282481 +0000 UTC m=+6104.863893506" watchObservedRunningTime="2025-12-10 20:37:04.365239541 +0000 UTC m=+6104.875850546" Dec 10 20:37:10 crc kubenswrapper[4828]: I1210 20:37:10.357144 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fq8zv" Dec 10 20:37:10 crc kubenswrapper[4828]: I1210 20:37:10.357838 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fq8zv" Dec 10 20:37:10 crc kubenswrapper[4828]: I1210 20:37:10.416334 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fq8zv" Dec 10 20:37:10 crc kubenswrapper[4828]: I1210 20:37:10.487883 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fq8zv" Dec 10 20:37:10 crc kubenswrapper[4828]: I1210 20:37:10.666176 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fq8zv"] Dec 10 20:37:12 crc kubenswrapper[4828]: I1210 20:37:12.417218 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fq8zv" podUID="3fefba56-6955-484a-a03f-7d46ffe45540" containerName="registry-server" containerID="cri-o://33054382e319c3e206537433bc39aa8e9f39d8ac4571f9f445f5e1baf7bc46b0" gracePeriod=2 Dec 10 20:37:12 crc kubenswrapper[4828]: I1210 20:37:12.933236 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fq8zv" Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.095917 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fefba56-6955-484a-a03f-7d46ffe45540-catalog-content\") pod \"3fefba56-6955-484a-a03f-7d46ffe45540\" (UID: \"3fefba56-6955-484a-a03f-7d46ffe45540\") " Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.096145 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fefba56-6955-484a-a03f-7d46ffe45540-utilities\") pod \"3fefba56-6955-484a-a03f-7d46ffe45540\" (UID: \"3fefba56-6955-484a-a03f-7d46ffe45540\") " Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.097033 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fefba56-6955-484a-a03f-7d46ffe45540-utilities" (OuterVolumeSpecName: "utilities") pod "3fefba56-6955-484a-a03f-7d46ffe45540" (UID: "3fefba56-6955-484a-a03f-7d46ffe45540"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.097162 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xntbw\" (UniqueName: \"kubernetes.io/projected/3fefba56-6955-484a-a03f-7d46ffe45540-kube-api-access-xntbw\") pod \"3fefba56-6955-484a-a03f-7d46ffe45540\" (UID: \"3fefba56-6955-484a-a03f-7d46ffe45540\") " Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.097661 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fefba56-6955-484a-a03f-7d46ffe45540-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.103412 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fefba56-6955-484a-a03f-7d46ffe45540-kube-api-access-xntbw" (OuterVolumeSpecName: "kube-api-access-xntbw") pod "3fefba56-6955-484a-a03f-7d46ffe45540" (UID: "3fefba56-6955-484a-a03f-7d46ffe45540"). InnerVolumeSpecName "kube-api-access-xntbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.143001 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fefba56-6955-484a-a03f-7d46ffe45540-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3fefba56-6955-484a-a03f-7d46ffe45540" (UID: "3fefba56-6955-484a-a03f-7d46ffe45540"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.199990 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fefba56-6955-484a-a03f-7d46ffe45540-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.200027 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xntbw\" (UniqueName: \"kubernetes.io/projected/3fefba56-6955-484a-a03f-7d46ffe45540-kube-api-access-xntbw\") on node \"crc\" DevicePath \"\"" Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.430676 4828 generic.go:334] "Generic (PLEG): container finished" podID="3fefba56-6955-484a-a03f-7d46ffe45540" containerID="33054382e319c3e206537433bc39aa8e9f39d8ac4571f9f445f5e1baf7bc46b0" exitCode=0 Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.430730 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fq8zv" event={"ID":"3fefba56-6955-484a-a03f-7d46ffe45540","Type":"ContainerDied","Data":"33054382e319c3e206537433bc39aa8e9f39d8ac4571f9f445f5e1baf7bc46b0"} Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.430761 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fq8zv" event={"ID":"3fefba56-6955-484a-a03f-7d46ffe45540","Type":"ContainerDied","Data":"ab15b656ac21c1c876f769b2fb34f30972058560e91387ec0fe5f491abef419b"} Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.430782 4828 scope.go:117] "RemoveContainer" containerID="33054382e319c3e206537433bc39aa8e9f39d8ac4571f9f445f5e1baf7bc46b0" Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.430967 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fq8zv" Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.471517 4828 scope.go:117] "RemoveContainer" containerID="28517f630bc0a5ec40ea949b3a2bac46e3d9b4de89386cd457579ff31cee70fb" Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.496106 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fq8zv"] Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.507396 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fq8zv"] Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.510230 4828 scope.go:117] "RemoveContainer" containerID="d1ed28c7d79f3b3b62f3d6133061e9417bfdc585a19fe277dedfb95cdd1e7fa6" Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.553168 4828 scope.go:117] "RemoveContainer" containerID="33054382e319c3e206537433bc39aa8e9f39d8ac4571f9f445f5e1baf7bc46b0" Dec 10 20:37:13 crc kubenswrapper[4828]: E1210 20:37:13.553694 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33054382e319c3e206537433bc39aa8e9f39d8ac4571f9f445f5e1baf7bc46b0\": container with ID starting with 33054382e319c3e206537433bc39aa8e9f39d8ac4571f9f445f5e1baf7bc46b0 not found: ID does not exist" containerID="33054382e319c3e206537433bc39aa8e9f39d8ac4571f9f445f5e1baf7bc46b0" Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.553727 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33054382e319c3e206537433bc39aa8e9f39d8ac4571f9f445f5e1baf7bc46b0"} err="failed to get container status \"33054382e319c3e206537433bc39aa8e9f39d8ac4571f9f445f5e1baf7bc46b0\": rpc error: code = NotFound desc = could not find container \"33054382e319c3e206537433bc39aa8e9f39d8ac4571f9f445f5e1baf7bc46b0\": container with ID starting with 33054382e319c3e206537433bc39aa8e9f39d8ac4571f9f445f5e1baf7bc46b0 not found: ID does not exist" Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.553748 4828 scope.go:117] "RemoveContainer" containerID="28517f630bc0a5ec40ea949b3a2bac46e3d9b4de89386cd457579ff31cee70fb" Dec 10 20:37:13 crc kubenswrapper[4828]: E1210 20:37:13.554043 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28517f630bc0a5ec40ea949b3a2bac46e3d9b4de89386cd457579ff31cee70fb\": container with ID starting with 28517f630bc0a5ec40ea949b3a2bac46e3d9b4de89386cd457579ff31cee70fb not found: ID does not exist" containerID="28517f630bc0a5ec40ea949b3a2bac46e3d9b4de89386cd457579ff31cee70fb" Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.554061 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28517f630bc0a5ec40ea949b3a2bac46e3d9b4de89386cd457579ff31cee70fb"} err="failed to get container status \"28517f630bc0a5ec40ea949b3a2bac46e3d9b4de89386cd457579ff31cee70fb\": rpc error: code = NotFound desc = could not find container \"28517f630bc0a5ec40ea949b3a2bac46e3d9b4de89386cd457579ff31cee70fb\": container with ID starting with 28517f630bc0a5ec40ea949b3a2bac46e3d9b4de89386cd457579ff31cee70fb not found: ID does not exist" Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.554075 4828 scope.go:117] "RemoveContainer" containerID="d1ed28c7d79f3b3b62f3d6133061e9417bfdc585a19fe277dedfb95cdd1e7fa6" Dec 10 20:37:13 crc kubenswrapper[4828]: E1210 20:37:13.554311 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1ed28c7d79f3b3b62f3d6133061e9417bfdc585a19fe277dedfb95cdd1e7fa6\": container with ID starting with d1ed28c7d79f3b3b62f3d6133061e9417bfdc585a19fe277dedfb95cdd1e7fa6 not found: ID does not exist" containerID="d1ed28c7d79f3b3b62f3d6133061e9417bfdc585a19fe277dedfb95cdd1e7fa6" Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.554329 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1ed28c7d79f3b3b62f3d6133061e9417bfdc585a19fe277dedfb95cdd1e7fa6"} err="failed to get container status \"d1ed28c7d79f3b3b62f3d6133061e9417bfdc585a19fe277dedfb95cdd1e7fa6\": rpc error: code = NotFound desc = could not find container \"d1ed28c7d79f3b3b62f3d6133061e9417bfdc585a19fe277dedfb95cdd1e7fa6\": container with ID starting with d1ed28c7d79f3b3b62f3d6133061e9417bfdc585a19fe277dedfb95cdd1e7fa6 not found: ID does not exist" Dec 10 20:37:13 crc kubenswrapper[4828]: I1210 20:37:13.801367 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fefba56-6955-484a-a03f-7d46ffe45540" path="/var/lib/kubelet/pods/3fefba56-6955-484a-a03f-7d46ffe45540/volumes" Dec 10 20:37:21 crc kubenswrapper[4828]: I1210 20:37:21.230420 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:37:21 crc kubenswrapper[4828]: I1210 20:37:21.232137 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:37:26 crc kubenswrapper[4828]: I1210 20:37:26.616891 4828 generic.go:334] "Generic (PLEG): container finished" podID="e4454add-ea32-4989-a3ef-700e057eed1a" containerID="0a1fe51129829cad3c79669a585e9268fbd56911b8d54f0f5e4d174b4743f60c" exitCode=0 Dec 10 20:37:26 crc kubenswrapper[4828]: I1210 20:37:26.617196 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qpwwn/must-gather-n5s68" event={"ID":"e4454add-ea32-4989-a3ef-700e057eed1a","Type":"ContainerDied","Data":"0a1fe51129829cad3c79669a585e9268fbd56911b8d54f0f5e4d174b4743f60c"} Dec 10 20:37:26 crc kubenswrapper[4828]: I1210 20:37:26.618981 4828 scope.go:117] "RemoveContainer" containerID="0a1fe51129829cad3c79669a585e9268fbd56911b8d54f0f5e4d174b4743f60c" Dec 10 20:37:26 crc kubenswrapper[4828]: I1210 20:37:26.706330 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qpwwn_must-gather-n5s68_e4454add-ea32-4989-a3ef-700e057eed1a/gather/0.log" Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.095938 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qpwwn/must-gather-n5s68"] Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.097045 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-qpwwn/must-gather-n5s68" podUID="e4454add-ea32-4989-a3ef-700e057eed1a" containerName="copy" containerID="cri-o://324ffea16ebb8fc04036ee280c3a99820b4d0b87426ffd21adb809e90a4d7f72" gracePeriod=2 Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.111611 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qpwwn/must-gather-n5s68"] Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.635778 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qpwwn_must-gather-n5s68_e4454add-ea32-4989-a3ef-700e057eed1a/copy/0.log" Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.636511 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qpwwn/must-gather-n5s68" Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.725000 4828 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qpwwn_must-gather-n5s68_e4454add-ea32-4989-a3ef-700e057eed1a/copy/0.log" Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.725667 4828 generic.go:334] "Generic (PLEG): container finished" podID="e4454add-ea32-4989-a3ef-700e057eed1a" containerID="324ffea16ebb8fc04036ee280c3a99820b4d0b87426ffd21adb809e90a4d7f72" exitCode=143 Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.725718 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qpwwn/must-gather-n5s68" Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.725812 4828 scope.go:117] "RemoveContainer" containerID="324ffea16ebb8fc04036ee280c3a99820b4d0b87426ffd21adb809e90a4d7f72" Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.746452 4828 scope.go:117] "RemoveContainer" containerID="0a1fe51129829cad3c79669a585e9268fbd56911b8d54f0f5e4d174b4743f60c" Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.788777 4828 scope.go:117] "RemoveContainer" containerID="324ffea16ebb8fc04036ee280c3a99820b4d0b87426ffd21adb809e90a4d7f72" Dec 10 20:37:36 crc kubenswrapper[4828]: E1210 20:37:36.790372 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"324ffea16ebb8fc04036ee280c3a99820b4d0b87426ffd21adb809e90a4d7f72\": container with ID starting with 324ffea16ebb8fc04036ee280c3a99820b4d0b87426ffd21adb809e90a4d7f72 not found: ID does not exist" containerID="324ffea16ebb8fc04036ee280c3a99820b4d0b87426ffd21adb809e90a4d7f72" Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.790433 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"324ffea16ebb8fc04036ee280c3a99820b4d0b87426ffd21adb809e90a4d7f72"} err="failed to get container status \"324ffea16ebb8fc04036ee280c3a99820b4d0b87426ffd21adb809e90a4d7f72\": rpc error: code = NotFound desc = could not find container \"324ffea16ebb8fc04036ee280c3a99820b4d0b87426ffd21adb809e90a4d7f72\": container with ID starting with 324ffea16ebb8fc04036ee280c3a99820b4d0b87426ffd21adb809e90a4d7f72 not found: ID does not exist" Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.790467 4828 scope.go:117] "RemoveContainer" containerID="0a1fe51129829cad3c79669a585e9268fbd56911b8d54f0f5e4d174b4743f60c" Dec 10 20:37:36 crc kubenswrapper[4828]: E1210 20:37:36.790780 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a1fe51129829cad3c79669a585e9268fbd56911b8d54f0f5e4d174b4743f60c\": container with ID starting with 0a1fe51129829cad3c79669a585e9268fbd56911b8d54f0f5e4d174b4743f60c not found: ID does not exist" containerID="0a1fe51129829cad3c79669a585e9268fbd56911b8d54f0f5e4d174b4743f60c" Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.790968 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a1fe51129829cad3c79669a585e9268fbd56911b8d54f0f5e4d174b4743f60c"} err="failed to get container status \"0a1fe51129829cad3c79669a585e9268fbd56911b8d54f0f5e4d174b4743f60c\": rpc error: code = NotFound desc = could not find container \"0a1fe51129829cad3c79669a585e9268fbd56911b8d54f0f5e4d174b4743f60c\": container with ID starting with 0a1fe51129829cad3c79669a585e9268fbd56911b8d54f0f5e4d174b4743f60c not found: ID does not exist" Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.807192 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e4454add-ea32-4989-a3ef-700e057eed1a-must-gather-output\") pod \"e4454add-ea32-4989-a3ef-700e057eed1a\" (UID: \"e4454add-ea32-4989-a3ef-700e057eed1a\") " Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.807621 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9pff\" (UniqueName: \"kubernetes.io/projected/e4454add-ea32-4989-a3ef-700e057eed1a-kube-api-access-k9pff\") pod \"e4454add-ea32-4989-a3ef-700e057eed1a\" (UID: \"e4454add-ea32-4989-a3ef-700e057eed1a\") " Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.814212 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4454add-ea32-4989-a3ef-700e057eed1a-kube-api-access-k9pff" (OuterVolumeSpecName: "kube-api-access-k9pff") pod "e4454add-ea32-4989-a3ef-700e057eed1a" (UID: "e4454add-ea32-4989-a3ef-700e057eed1a"). InnerVolumeSpecName "kube-api-access-k9pff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:37:36 crc kubenswrapper[4828]: I1210 20:37:36.910897 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9pff\" (UniqueName: \"kubernetes.io/projected/e4454add-ea32-4989-a3ef-700e057eed1a-kube-api-access-k9pff\") on node \"crc\" DevicePath \"\"" Dec 10 20:37:37 crc kubenswrapper[4828]: I1210 20:37:37.018628 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4454add-ea32-4989-a3ef-700e057eed1a-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e4454add-ea32-4989-a3ef-700e057eed1a" (UID: "e4454add-ea32-4989-a3ef-700e057eed1a"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:37:37 crc kubenswrapper[4828]: I1210 20:37:37.115329 4828 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e4454add-ea32-4989-a3ef-700e057eed1a-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 10 20:37:37 crc kubenswrapper[4828]: I1210 20:37:37.800891 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4454add-ea32-4989-a3ef-700e057eed1a" path="/var/lib/kubelet/pods/e4454add-ea32-4989-a3ef-700e057eed1a/volumes" Dec 10 20:37:45 crc kubenswrapper[4828]: I1210 20:37:45.422484 4828 scope.go:117] "RemoveContainer" containerID="cd93cc78c51aae8e287b9abfebc85e1f6c9880808a1dbb8e44e760a74eb3aa12" Dec 10 20:37:51 crc kubenswrapper[4828]: I1210 20:37:51.230616 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:37:51 crc kubenswrapper[4828]: I1210 20:37:51.231279 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.035099 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8tb8b"] Dec 10 20:38:02 crc kubenswrapper[4828]: E1210 20:38:02.036332 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fefba56-6955-484a-a03f-7d46ffe45540" containerName="extract-content" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.036350 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fefba56-6955-484a-a03f-7d46ffe45540" containerName="extract-content" Dec 10 20:38:02 crc kubenswrapper[4828]: E1210 20:38:02.036370 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4454add-ea32-4989-a3ef-700e057eed1a" containerName="copy" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.036381 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4454add-ea32-4989-a3ef-700e057eed1a" containerName="copy" Dec 10 20:38:02 crc kubenswrapper[4828]: E1210 20:38:02.036406 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fefba56-6955-484a-a03f-7d46ffe45540" containerName="extract-utilities" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.036415 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fefba56-6955-484a-a03f-7d46ffe45540" containerName="extract-utilities" Dec 10 20:38:02 crc kubenswrapper[4828]: E1210 20:38:02.036429 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fefba56-6955-484a-a03f-7d46ffe45540" containerName="registry-server" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.036437 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fefba56-6955-484a-a03f-7d46ffe45540" containerName="registry-server" Dec 10 20:38:02 crc kubenswrapper[4828]: E1210 20:38:02.036460 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4454add-ea32-4989-a3ef-700e057eed1a" containerName="gather" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.036471 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4454add-ea32-4989-a3ef-700e057eed1a" containerName="gather" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.036767 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4454add-ea32-4989-a3ef-700e057eed1a" containerName="copy" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.036822 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4454add-ea32-4989-a3ef-700e057eed1a" containerName="gather" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.036842 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fefba56-6955-484a-a03f-7d46ffe45540" containerName="registry-server" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.039833 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8tb8b" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.046954 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8tb8b"] Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.076466 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q49v\" (UniqueName: \"kubernetes.io/projected/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-kube-api-access-5q49v\") pod \"community-operators-8tb8b\" (UID: \"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff\") " pod="openshift-marketplace/community-operators-8tb8b" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.076768 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-catalog-content\") pod \"community-operators-8tb8b\" (UID: \"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff\") " pod="openshift-marketplace/community-operators-8tb8b" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.076832 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-utilities\") pod \"community-operators-8tb8b\" (UID: \"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff\") " pod="openshift-marketplace/community-operators-8tb8b" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.178819 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-catalog-content\") pod \"community-operators-8tb8b\" (UID: \"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff\") " pod="openshift-marketplace/community-operators-8tb8b" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.178886 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-utilities\") pod \"community-operators-8tb8b\" (UID: \"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff\") " pod="openshift-marketplace/community-operators-8tb8b" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.178961 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q49v\" (UniqueName: \"kubernetes.io/projected/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-kube-api-access-5q49v\") pod \"community-operators-8tb8b\" (UID: \"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff\") " pod="openshift-marketplace/community-operators-8tb8b" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.179575 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-catalog-content\") pod \"community-operators-8tb8b\" (UID: \"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff\") " pod="openshift-marketplace/community-operators-8tb8b" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.179827 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-utilities\") pod \"community-operators-8tb8b\" (UID: \"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff\") " pod="openshift-marketplace/community-operators-8tb8b" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.202766 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q49v\" (UniqueName: \"kubernetes.io/projected/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-kube-api-access-5q49v\") pod \"community-operators-8tb8b\" (UID: \"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff\") " pod="openshift-marketplace/community-operators-8tb8b" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.375995 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8tb8b" Dec 10 20:38:02 crc kubenswrapper[4828]: I1210 20:38:02.926377 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8tb8b"] Dec 10 20:38:03 crc kubenswrapper[4828]: I1210 20:38:03.018714 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8tb8b" event={"ID":"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff","Type":"ContainerStarted","Data":"79133e57c3a7d6f3fb0043dd76e3c8713b472f76b95191b587eb9d6e48007c03"} Dec 10 20:38:04 crc kubenswrapper[4828]: I1210 20:38:04.041865 4828 generic.go:334] "Generic (PLEG): container finished" podID="9aefc5d3-9e7f-4f06-93e1-22cbb03bddff" containerID="739f6d2f319a48f48a32f1d1f0b3349da659b74ba55613504016807740aa7069" exitCode=0 Dec 10 20:38:04 crc kubenswrapper[4828]: I1210 20:38:04.042701 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8tb8b" event={"ID":"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff","Type":"ContainerDied","Data":"739f6d2f319a48f48a32f1d1f0b3349da659b74ba55613504016807740aa7069"} Dec 10 20:38:06 crc kubenswrapper[4828]: I1210 20:38:06.066511 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8tb8b" event={"ID":"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff","Type":"ContainerStarted","Data":"7ee012868304af048e9140457d0e49740eba334a92df016d9f269c4abf639468"} Dec 10 20:38:07 crc kubenswrapper[4828]: I1210 20:38:07.083289 4828 generic.go:334] "Generic (PLEG): container finished" podID="9aefc5d3-9e7f-4f06-93e1-22cbb03bddff" containerID="7ee012868304af048e9140457d0e49740eba334a92df016d9f269c4abf639468" exitCode=0 Dec 10 20:38:07 crc kubenswrapper[4828]: I1210 20:38:07.083378 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8tb8b" event={"ID":"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff","Type":"ContainerDied","Data":"7ee012868304af048e9140457d0e49740eba334a92df016d9f269c4abf639468"} Dec 10 20:38:09 crc kubenswrapper[4828]: I1210 20:38:09.129967 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8tb8b" event={"ID":"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff","Type":"ContainerStarted","Data":"4e994ca2a88fa75743be9373f49ebb58b0668174cf5d2fbc05d547b7c0ce723c"} Dec 10 20:38:09 crc kubenswrapper[4828]: I1210 20:38:09.160236 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8tb8b" podStartSLOduration=3.254144263 podStartE2EDuration="7.160218193s" podCreationTimestamp="2025-12-10 20:38:02 +0000 UTC" firstStartedPulling="2025-12-10 20:38:04.046241354 +0000 UTC m=+6164.556852359" lastFinishedPulling="2025-12-10 20:38:07.952315284 +0000 UTC m=+6168.462926289" observedRunningTime="2025-12-10 20:38:09.159569646 +0000 UTC m=+6169.670180651" watchObservedRunningTime="2025-12-10 20:38:09.160218193 +0000 UTC m=+6169.670829188" Dec 10 20:38:12 crc kubenswrapper[4828]: I1210 20:38:12.376459 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8tb8b" Dec 10 20:38:12 crc kubenswrapper[4828]: I1210 20:38:12.377117 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8tb8b" Dec 10 20:38:12 crc kubenswrapper[4828]: I1210 20:38:12.428930 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8tb8b" Dec 10 20:38:13 crc kubenswrapper[4828]: I1210 20:38:13.264341 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8tb8b" Dec 10 20:38:13 crc kubenswrapper[4828]: I1210 20:38:13.319993 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8tb8b"] Dec 10 20:38:15 crc kubenswrapper[4828]: I1210 20:38:15.201263 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8tb8b" podUID="9aefc5d3-9e7f-4f06-93e1-22cbb03bddff" containerName="registry-server" containerID="cri-o://4e994ca2a88fa75743be9373f49ebb58b0668174cf5d2fbc05d547b7c0ce723c" gracePeriod=2 Dec 10 20:38:15 crc kubenswrapper[4828]: I1210 20:38:15.702597 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8tb8b" Dec 10 20:38:15 crc kubenswrapper[4828]: I1210 20:38:15.805035 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5q49v\" (UniqueName: \"kubernetes.io/projected/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-kube-api-access-5q49v\") pod \"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff\" (UID: \"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff\") " Dec 10 20:38:15 crc kubenswrapper[4828]: I1210 20:38:15.805151 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-catalog-content\") pod \"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff\" (UID: \"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff\") " Dec 10 20:38:15 crc kubenswrapper[4828]: I1210 20:38:15.805442 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-utilities\") pod \"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff\" (UID: \"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff\") " Dec 10 20:38:15 crc kubenswrapper[4828]: I1210 20:38:15.806774 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-utilities" (OuterVolumeSpecName: "utilities") pod "9aefc5d3-9e7f-4f06-93e1-22cbb03bddff" (UID: "9aefc5d3-9e7f-4f06-93e1-22cbb03bddff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:38:15 crc kubenswrapper[4828]: I1210 20:38:15.811434 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-kube-api-access-5q49v" (OuterVolumeSpecName: "kube-api-access-5q49v") pod "9aefc5d3-9e7f-4f06-93e1-22cbb03bddff" (UID: "9aefc5d3-9e7f-4f06-93e1-22cbb03bddff"). InnerVolumeSpecName "kube-api-access-5q49v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:38:15 crc kubenswrapper[4828]: I1210 20:38:15.849691 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9aefc5d3-9e7f-4f06-93e1-22cbb03bddff" (UID: "9aefc5d3-9e7f-4f06-93e1-22cbb03bddff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:38:15 crc kubenswrapper[4828]: I1210 20:38:15.908366 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:38:15 crc kubenswrapper[4828]: I1210 20:38:15.908410 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5q49v\" (UniqueName: \"kubernetes.io/projected/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-kube-api-access-5q49v\") on node \"crc\" DevicePath \"\"" Dec 10 20:38:15 crc kubenswrapper[4828]: I1210 20:38:15.908428 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:38:16 crc kubenswrapper[4828]: I1210 20:38:16.213711 4828 generic.go:334] "Generic (PLEG): container finished" podID="9aefc5d3-9e7f-4f06-93e1-22cbb03bddff" containerID="4e994ca2a88fa75743be9373f49ebb58b0668174cf5d2fbc05d547b7c0ce723c" exitCode=0 Dec 10 20:38:16 crc kubenswrapper[4828]: I1210 20:38:16.213762 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8tb8b" event={"ID":"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff","Type":"ContainerDied","Data":"4e994ca2a88fa75743be9373f49ebb58b0668174cf5d2fbc05d547b7c0ce723c"} Dec 10 20:38:16 crc kubenswrapper[4828]: I1210 20:38:16.213787 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8tb8b" Dec 10 20:38:16 crc kubenswrapper[4828]: I1210 20:38:16.213831 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8tb8b" event={"ID":"9aefc5d3-9e7f-4f06-93e1-22cbb03bddff","Type":"ContainerDied","Data":"79133e57c3a7d6f3fb0043dd76e3c8713b472f76b95191b587eb9d6e48007c03"} Dec 10 20:38:16 crc kubenswrapper[4828]: I1210 20:38:16.213889 4828 scope.go:117] "RemoveContainer" containerID="4e994ca2a88fa75743be9373f49ebb58b0668174cf5d2fbc05d547b7c0ce723c" Dec 10 20:38:16 crc kubenswrapper[4828]: I1210 20:38:16.240871 4828 scope.go:117] "RemoveContainer" containerID="7ee012868304af048e9140457d0e49740eba334a92df016d9f269c4abf639468" Dec 10 20:38:16 crc kubenswrapper[4828]: I1210 20:38:16.248937 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8tb8b"] Dec 10 20:38:16 crc kubenswrapper[4828]: I1210 20:38:16.261097 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8tb8b"] Dec 10 20:38:16 crc kubenswrapper[4828]: I1210 20:38:16.265787 4828 scope.go:117] "RemoveContainer" containerID="739f6d2f319a48f48a32f1d1f0b3349da659b74ba55613504016807740aa7069" Dec 10 20:38:16 crc kubenswrapper[4828]: I1210 20:38:16.333656 4828 scope.go:117] "RemoveContainer" containerID="4e994ca2a88fa75743be9373f49ebb58b0668174cf5d2fbc05d547b7c0ce723c" Dec 10 20:38:16 crc kubenswrapper[4828]: E1210 20:38:16.334151 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e994ca2a88fa75743be9373f49ebb58b0668174cf5d2fbc05d547b7c0ce723c\": container with ID starting with 4e994ca2a88fa75743be9373f49ebb58b0668174cf5d2fbc05d547b7c0ce723c not found: ID does not exist" containerID="4e994ca2a88fa75743be9373f49ebb58b0668174cf5d2fbc05d547b7c0ce723c" Dec 10 20:38:16 crc kubenswrapper[4828]: I1210 20:38:16.334195 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e994ca2a88fa75743be9373f49ebb58b0668174cf5d2fbc05d547b7c0ce723c"} err="failed to get container status \"4e994ca2a88fa75743be9373f49ebb58b0668174cf5d2fbc05d547b7c0ce723c\": rpc error: code = NotFound desc = could not find container \"4e994ca2a88fa75743be9373f49ebb58b0668174cf5d2fbc05d547b7c0ce723c\": container with ID starting with 4e994ca2a88fa75743be9373f49ebb58b0668174cf5d2fbc05d547b7c0ce723c not found: ID does not exist" Dec 10 20:38:16 crc kubenswrapper[4828]: I1210 20:38:16.334223 4828 scope.go:117] "RemoveContainer" containerID="7ee012868304af048e9140457d0e49740eba334a92df016d9f269c4abf639468" Dec 10 20:38:16 crc kubenswrapper[4828]: E1210 20:38:16.334533 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ee012868304af048e9140457d0e49740eba334a92df016d9f269c4abf639468\": container with ID starting with 7ee012868304af048e9140457d0e49740eba334a92df016d9f269c4abf639468 not found: ID does not exist" containerID="7ee012868304af048e9140457d0e49740eba334a92df016d9f269c4abf639468" Dec 10 20:38:16 crc kubenswrapper[4828]: I1210 20:38:16.334577 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ee012868304af048e9140457d0e49740eba334a92df016d9f269c4abf639468"} err="failed to get container status \"7ee012868304af048e9140457d0e49740eba334a92df016d9f269c4abf639468\": rpc error: code = NotFound desc = could not find container \"7ee012868304af048e9140457d0e49740eba334a92df016d9f269c4abf639468\": container with ID starting with 7ee012868304af048e9140457d0e49740eba334a92df016d9f269c4abf639468 not found: ID does not exist" Dec 10 20:38:16 crc kubenswrapper[4828]: I1210 20:38:16.334605 4828 scope.go:117] "RemoveContainer" containerID="739f6d2f319a48f48a32f1d1f0b3349da659b74ba55613504016807740aa7069" Dec 10 20:38:16 crc kubenswrapper[4828]: E1210 20:38:16.334901 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"739f6d2f319a48f48a32f1d1f0b3349da659b74ba55613504016807740aa7069\": container with ID starting with 739f6d2f319a48f48a32f1d1f0b3349da659b74ba55613504016807740aa7069 not found: ID does not exist" containerID="739f6d2f319a48f48a32f1d1f0b3349da659b74ba55613504016807740aa7069" Dec 10 20:38:16 crc kubenswrapper[4828]: I1210 20:38:16.334928 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"739f6d2f319a48f48a32f1d1f0b3349da659b74ba55613504016807740aa7069"} err="failed to get container status \"739f6d2f319a48f48a32f1d1f0b3349da659b74ba55613504016807740aa7069\": rpc error: code = NotFound desc = could not find container \"739f6d2f319a48f48a32f1d1f0b3349da659b74ba55613504016807740aa7069\": container with ID starting with 739f6d2f319a48f48a32f1d1f0b3349da659b74ba55613504016807740aa7069 not found: ID does not exist" Dec 10 20:38:17 crc kubenswrapper[4828]: I1210 20:38:17.810490 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aefc5d3-9e7f-4f06-93e1-22cbb03bddff" path="/var/lib/kubelet/pods/9aefc5d3-9e7f-4f06-93e1-22cbb03bddff/volumes" Dec 10 20:38:21 crc kubenswrapper[4828]: I1210 20:38:21.230391 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:38:21 crc kubenswrapper[4828]: I1210 20:38:21.230917 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:38:21 crc kubenswrapper[4828]: I1210 20:38:21.230957 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 20:38:21 crc kubenswrapper[4828]: I1210 20:38:21.231707 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"325ae6fd60148173e9b100cf5cb87a1e61d379c9db65348aeb2ae47852a05290"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 20:38:21 crc kubenswrapper[4828]: I1210 20:38:21.231757 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://325ae6fd60148173e9b100cf5cb87a1e61d379c9db65348aeb2ae47852a05290" gracePeriod=600 Dec 10 20:38:22 crc kubenswrapper[4828]: I1210 20:38:22.280883 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="325ae6fd60148173e9b100cf5cb87a1e61d379c9db65348aeb2ae47852a05290" exitCode=0 Dec 10 20:38:22 crc kubenswrapper[4828]: I1210 20:38:22.280930 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"325ae6fd60148173e9b100cf5cb87a1e61d379c9db65348aeb2ae47852a05290"} Dec 10 20:38:22 crc kubenswrapper[4828]: I1210 20:38:22.281198 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerStarted","Data":"039316a83fc8a6974ee10f5ecdceec1e0fbd066ac42e73d4bf48cef8444d9fe2"} Dec 10 20:38:22 crc kubenswrapper[4828]: I1210 20:38:22.281220 4828 scope.go:117] "RemoveContainer" containerID="577486f385b81df406d5c76e8e52cd1329c2db1d05440a24aac6e6b46ef565b2" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.380875 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fqg8n"] Dec 10 20:38:35 crc kubenswrapper[4828]: E1210 20:38:35.382350 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aefc5d3-9e7f-4f06-93e1-22cbb03bddff" containerName="extract-utilities" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.382375 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aefc5d3-9e7f-4f06-93e1-22cbb03bddff" containerName="extract-utilities" Dec 10 20:38:35 crc kubenswrapper[4828]: E1210 20:38:35.382417 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aefc5d3-9e7f-4f06-93e1-22cbb03bddff" containerName="extract-content" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.382430 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aefc5d3-9e7f-4f06-93e1-22cbb03bddff" containerName="extract-content" Dec 10 20:38:35 crc kubenswrapper[4828]: E1210 20:38:35.382462 4828 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aefc5d3-9e7f-4f06-93e1-22cbb03bddff" containerName="registry-server" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.382473 4828 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aefc5d3-9e7f-4f06-93e1-22cbb03bddff" containerName="registry-server" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.382859 4828 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aefc5d3-9e7f-4f06-93e1-22cbb03bddff" containerName="registry-server" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.385780 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fqg8n" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.401088 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqg8n"] Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.577409 4828 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-95vqh"] Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.580221 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-95vqh" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.591343 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9013fd72-548b-4f6b-9e6e-e0fdd9541900-catalog-content\") pod \"redhat-operators-95vqh\" (UID: \"9013fd72-548b-4f6b-9e6e-e0fdd9541900\") " pod="openshift-marketplace/redhat-operators-95vqh" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.591663 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-catalog-content\") pod \"redhat-marketplace-fqg8n\" (UID: \"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8\") " pod="openshift-marketplace/redhat-marketplace-fqg8n" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.591771 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zndbr\" (UniqueName: \"kubernetes.io/projected/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-kube-api-access-zndbr\") pod \"redhat-marketplace-fqg8n\" (UID: \"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8\") " pod="openshift-marketplace/redhat-marketplace-fqg8n" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.591904 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9013fd72-548b-4f6b-9e6e-e0fdd9541900-utilities\") pod \"redhat-operators-95vqh\" (UID: \"9013fd72-548b-4f6b-9e6e-e0fdd9541900\") " pod="openshift-marketplace/redhat-operators-95vqh" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.592198 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncj2p\" (UniqueName: \"kubernetes.io/projected/9013fd72-548b-4f6b-9e6e-e0fdd9541900-kube-api-access-ncj2p\") pod \"redhat-operators-95vqh\" (UID: \"9013fd72-548b-4f6b-9e6e-e0fdd9541900\") " pod="openshift-marketplace/redhat-operators-95vqh" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.592378 4828 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-utilities\") pod \"redhat-marketplace-fqg8n\" (UID: \"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8\") " pod="openshift-marketplace/redhat-marketplace-fqg8n" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.618084 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-95vqh"] Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.694500 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncj2p\" (UniqueName: \"kubernetes.io/projected/9013fd72-548b-4f6b-9e6e-e0fdd9541900-kube-api-access-ncj2p\") pod \"redhat-operators-95vqh\" (UID: \"9013fd72-548b-4f6b-9e6e-e0fdd9541900\") " pod="openshift-marketplace/redhat-operators-95vqh" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.694584 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-utilities\") pod \"redhat-marketplace-fqg8n\" (UID: \"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8\") " pod="openshift-marketplace/redhat-marketplace-fqg8n" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.694674 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9013fd72-548b-4f6b-9e6e-e0fdd9541900-catalog-content\") pod \"redhat-operators-95vqh\" (UID: \"9013fd72-548b-4f6b-9e6e-e0fdd9541900\") " pod="openshift-marketplace/redhat-operators-95vqh" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.694708 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-catalog-content\") pod \"redhat-marketplace-fqg8n\" (UID: \"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8\") " pod="openshift-marketplace/redhat-marketplace-fqg8n" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.694730 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zndbr\" (UniqueName: \"kubernetes.io/projected/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-kube-api-access-zndbr\") pod \"redhat-marketplace-fqg8n\" (UID: \"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8\") " pod="openshift-marketplace/redhat-marketplace-fqg8n" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.694759 4828 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9013fd72-548b-4f6b-9e6e-e0fdd9541900-utilities\") pod \"redhat-operators-95vqh\" (UID: \"9013fd72-548b-4f6b-9e6e-e0fdd9541900\") " pod="openshift-marketplace/redhat-operators-95vqh" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.695187 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-utilities\") pod \"redhat-marketplace-fqg8n\" (UID: \"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8\") " pod="openshift-marketplace/redhat-marketplace-fqg8n" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.695301 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9013fd72-548b-4f6b-9e6e-e0fdd9541900-catalog-content\") pod \"redhat-operators-95vqh\" (UID: \"9013fd72-548b-4f6b-9e6e-e0fdd9541900\") " pod="openshift-marketplace/redhat-operators-95vqh" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.695387 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9013fd72-548b-4f6b-9e6e-e0fdd9541900-utilities\") pod \"redhat-operators-95vqh\" (UID: \"9013fd72-548b-4f6b-9e6e-e0fdd9541900\") " pod="openshift-marketplace/redhat-operators-95vqh" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.695486 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-catalog-content\") pod \"redhat-marketplace-fqg8n\" (UID: \"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8\") " pod="openshift-marketplace/redhat-marketplace-fqg8n" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.715563 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncj2p\" (UniqueName: \"kubernetes.io/projected/9013fd72-548b-4f6b-9e6e-e0fdd9541900-kube-api-access-ncj2p\") pod \"redhat-operators-95vqh\" (UID: \"9013fd72-548b-4f6b-9e6e-e0fdd9541900\") " pod="openshift-marketplace/redhat-operators-95vqh" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.717890 4828 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zndbr\" (UniqueName: \"kubernetes.io/projected/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-kube-api-access-zndbr\") pod \"redhat-marketplace-fqg8n\" (UID: \"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8\") " pod="openshift-marketplace/redhat-marketplace-fqg8n" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.721007 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fqg8n" Dec 10 20:38:35 crc kubenswrapper[4828]: I1210 20:38:35.907430 4828 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-95vqh" Dec 10 20:38:36 crc kubenswrapper[4828]: I1210 20:38:36.200295 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqg8n"] Dec 10 20:38:36 crc kubenswrapper[4828]: W1210 20:38:36.407719 4828 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9013fd72_548b_4f6b_9e6e_e0fdd9541900.slice/crio-066fcdf72d5cc0f655133d186e3f944c791067debd08b5fae07f74bf5cec886a WatchSource:0}: Error finding container 066fcdf72d5cc0f655133d186e3f944c791067debd08b5fae07f74bf5cec886a: Status 404 returned error can't find the container with id 066fcdf72d5cc0f655133d186e3f944c791067debd08b5fae07f74bf5cec886a Dec 10 20:38:36 crc kubenswrapper[4828]: I1210 20:38:36.416055 4828 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-95vqh"] Dec 10 20:38:36 crc kubenswrapper[4828]: I1210 20:38:36.436299 4828 generic.go:334] "Generic (PLEG): container finished" podID="2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8" containerID="e89bf57bf4777d6254ee68f5aacc94067c52005dbfca38f1318de3de79700b0a" exitCode=0 Dec 10 20:38:36 crc kubenswrapper[4828]: I1210 20:38:36.436386 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqg8n" event={"ID":"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8","Type":"ContainerDied","Data":"e89bf57bf4777d6254ee68f5aacc94067c52005dbfca38f1318de3de79700b0a"} Dec 10 20:38:36 crc kubenswrapper[4828]: I1210 20:38:36.436420 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqg8n" event={"ID":"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8","Type":"ContainerStarted","Data":"ef1ee93f5032aebbe2b89f27f94810dc47de8f4435cd460283cfe24110e87704"} Dec 10 20:38:36 crc kubenswrapper[4828]: I1210 20:38:36.437831 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95vqh" event={"ID":"9013fd72-548b-4f6b-9e6e-e0fdd9541900","Type":"ContainerStarted","Data":"066fcdf72d5cc0f655133d186e3f944c791067debd08b5fae07f74bf5cec886a"} Dec 10 20:38:37 crc kubenswrapper[4828]: I1210 20:38:37.451971 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqg8n" event={"ID":"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8","Type":"ContainerStarted","Data":"595a5aa7d2ce20a033806f8b57bc55abdcc874db342789ff09fac94113b9ab34"} Dec 10 20:38:37 crc kubenswrapper[4828]: I1210 20:38:37.456101 4828 generic.go:334] "Generic (PLEG): container finished" podID="9013fd72-548b-4f6b-9e6e-e0fdd9541900" containerID="6d701ad72875fbbeaddbed8af2aca59f216f62d35808423f7d44a61ebc21360d" exitCode=0 Dec 10 20:38:37 crc kubenswrapper[4828]: I1210 20:38:37.456156 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95vqh" event={"ID":"9013fd72-548b-4f6b-9e6e-e0fdd9541900","Type":"ContainerDied","Data":"6d701ad72875fbbeaddbed8af2aca59f216f62d35808423f7d44a61ebc21360d"} Dec 10 20:38:38 crc kubenswrapper[4828]: I1210 20:38:38.469080 4828 generic.go:334] "Generic (PLEG): container finished" podID="2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8" containerID="595a5aa7d2ce20a033806f8b57bc55abdcc874db342789ff09fac94113b9ab34" exitCode=0 Dec 10 20:38:38 crc kubenswrapper[4828]: I1210 20:38:38.469154 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqg8n" event={"ID":"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8","Type":"ContainerDied","Data":"595a5aa7d2ce20a033806f8b57bc55abdcc874db342789ff09fac94113b9ab34"} Dec 10 20:38:38 crc kubenswrapper[4828]: I1210 20:38:38.472451 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95vqh" event={"ID":"9013fd72-548b-4f6b-9e6e-e0fdd9541900","Type":"ContainerStarted","Data":"6fb81de34afd424e41abe011dd5712c4ef67e8b901683b256536cb08998fe53a"} Dec 10 20:38:41 crc kubenswrapper[4828]: I1210 20:38:41.504759 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqg8n" event={"ID":"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8","Type":"ContainerStarted","Data":"52487af24567560241ea250bbf51a155ea8692c8ed9ba88f0b12062861545e03"} Dec 10 20:38:41 crc kubenswrapper[4828]: I1210 20:38:41.525475 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fqg8n" podStartSLOduration=2.451013686 podStartE2EDuration="6.525457571s" podCreationTimestamp="2025-12-10 20:38:35 +0000 UTC" firstStartedPulling="2025-12-10 20:38:36.439676927 +0000 UTC m=+6196.950287932" lastFinishedPulling="2025-12-10 20:38:40.514120812 +0000 UTC m=+6201.024731817" observedRunningTime="2025-12-10 20:38:41.52427078 +0000 UTC m=+6202.034881795" watchObservedRunningTime="2025-12-10 20:38:41.525457571 +0000 UTC m=+6202.036068566" Dec 10 20:38:43 crc kubenswrapper[4828]: I1210 20:38:43.527523 4828 generic.go:334] "Generic (PLEG): container finished" podID="9013fd72-548b-4f6b-9e6e-e0fdd9541900" containerID="6fb81de34afd424e41abe011dd5712c4ef67e8b901683b256536cb08998fe53a" exitCode=0 Dec 10 20:38:43 crc kubenswrapper[4828]: I1210 20:38:43.527587 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95vqh" event={"ID":"9013fd72-548b-4f6b-9e6e-e0fdd9541900","Type":"ContainerDied","Data":"6fb81de34afd424e41abe011dd5712c4ef67e8b901683b256536cb08998fe53a"} Dec 10 20:38:44 crc kubenswrapper[4828]: I1210 20:38:44.541320 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95vqh" event={"ID":"9013fd72-548b-4f6b-9e6e-e0fdd9541900","Type":"ContainerStarted","Data":"4520def2eb44b069057a92cb95ae11e8936607be486b7a582559cadeaca841b5"} Dec 10 20:38:44 crc kubenswrapper[4828]: I1210 20:38:44.570596 4828 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-95vqh" podStartSLOduration=3.015045368 podStartE2EDuration="9.570572786s" podCreationTimestamp="2025-12-10 20:38:35 +0000 UTC" firstStartedPulling="2025-12-10 20:38:37.458193918 +0000 UTC m=+6197.968804923" lastFinishedPulling="2025-12-10 20:38:44.013721336 +0000 UTC m=+6204.524332341" observedRunningTime="2025-12-10 20:38:44.558487242 +0000 UTC m=+6205.069098237" watchObservedRunningTime="2025-12-10 20:38:44.570572786 +0000 UTC m=+6205.081183791" Dec 10 20:38:45 crc kubenswrapper[4828]: I1210 20:38:45.518711 4828 scope.go:117] "RemoveContainer" containerID="d7ab96cf13291f0ca1e526f6b65a4f4ae38ee562512b6d4e6b49de28c66d1ccf" Dec 10 20:38:45 crc kubenswrapper[4828]: I1210 20:38:45.722269 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fqg8n" Dec 10 20:38:45 crc kubenswrapper[4828]: I1210 20:38:45.722787 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fqg8n" Dec 10 20:38:45 crc kubenswrapper[4828]: I1210 20:38:45.772141 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fqg8n" Dec 10 20:38:45 crc kubenswrapper[4828]: I1210 20:38:45.907963 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-95vqh" Dec 10 20:38:45 crc kubenswrapper[4828]: I1210 20:38:45.908322 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-95vqh" Dec 10 20:38:46 crc kubenswrapper[4828]: I1210 20:38:46.615429 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fqg8n" Dec 10 20:38:46 crc kubenswrapper[4828]: I1210 20:38:46.959664 4828 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-95vqh" podUID="9013fd72-548b-4f6b-9e6e-e0fdd9541900" containerName="registry-server" probeResult="failure" output=< Dec 10 20:38:46 crc kubenswrapper[4828]: timeout: failed to connect service ":50051" within 1s Dec 10 20:38:46 crc kubenswrapper[4828]: > Dec 10 20:38:47 crc kubenswrapper[4828]: I1210 20:38:47.160442 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqg8n"] Dec 10 20:38:48 crc kubenswrapper[4828]: I1210 20:38:48.586688 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fqg8n" podUID="2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8" containerName="registry-server" containerID="cri-o://52487af24567560241ea250bbf51a155ea8692c8ed9ba88f0b12062861545e03" gracePeriod=2 Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.082923 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fqg8n" Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.233024 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zndbr\" (UniqueName: \"kubernetes.io/projected/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-kube-api-access-zndbr\") pod \"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8\" (UID: \"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8\") " Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.233435 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-catalog-content\") pod \"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8\" (UID: \"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8\") " Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.233561 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-utilities\") pod \"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8\" (UID: \"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8\") " Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.233969 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-utilities" (OuterVolumeSpecName: "utilities") pod "2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8" (UID: "2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.234341 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.239498 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-kube-api-access-zndbr" (OuterVolumeSpecName: "kube-api-access-zndbr") pod "2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8" (UID: "2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8"). InnerVolumeSpecName "kube-api-access-zndbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.254503 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8" (UID: "2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.336873 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zndbr\" (UniqueName: \"kubernetes.io/projected/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-kube-api-access-zndbr\") on node \"crc\" DevicePath \"\"" Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.337238 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.601598 4828 generic.go:334] "Generic (PLEG): container finished" podID="2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8" containerID="52487af24567560241ea250bbf51a155ea8692c8ed9ba88f0b12062861545e03" exitCode=0 Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.601642 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqg8n" event={"ID":"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8","Type":"ContainerDied","Data":"52487af24567560241ea250bbf51a155ea8692c8ed9ba88f0b12062861545e03"} Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.601662 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fqg8n" Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.601672 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqg8n" event={"ID":"2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8","Type":"ContainerDied","Data":"ef1ee93f5032aebbe2b89f27f94810dc47de8f4435cd460283cfe24110e87704"} Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.601688 4828 scope.go:117] "RemoveContainer" containerID="52487af24567560241ea250bbf51a155ea8692c8ed9ba88f0b12062861545e03" Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.625979 4828 scope.go:117] "RemoveContainer" containerID="595a5aa7d2ce20a033806f8b57bc55abdcc874db342789ff09fac94113b9ab34" Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.653563 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqg8n"] Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.661981 4828 scope.go:117] "RemoveContainer" containerID="e89bf57bf4777d6254ee68f5aacc94067c52005dbfca38f1318de3de79700b0a" Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.670716 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqg8n"] Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.720979 4828 scope.go:117] "RemoveContainer" containerID="52487af24567560241ea250bbf51a155ea8692c8ed9ba88f0b12062861545e03" Dec 10 20:38:49 crc kubenswrapper[4828]: E1210 20:38:49.721648 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52487af24567560241ea250bbf51a155ea8692c8ed9ba88f0b12062861545e03\": container with ID starting with 52487af24567560241ea250bbf51a155ea8692c8ed9ba88f0b12062861545e03 not found: ID does not exist" containerID="52487af24567560241ea250bbf51a155ea8692c8ed9ba88f0b12062861545e03" Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.721685 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52487af24567560241ea250bbf51a155ea8692c8ed9ba88f0b12062861545e03"} err="failed to get container status \"52487af24567560241ea250bbf51a155ea8692c8ed9ba88f0b12062861545e03\": rpc error: code = NotFound desc = could not find container \"52487af24567560241ea250bbf51a155ea8692c8ed9ba88f0b12062861545e03\": container with ID starting with 52487af24567560241ea250bbf51a155ea8692c8ed9ba88f0b12062861545e03 not found: ID does not exist" Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.721711 4828 scope.go:117] "RemoveContainer" containerID="595a5aa7d2ce20a033806f8b57bc55abdcc874db342789ff09fac94113b9ab34" Dec 10 20:38:49 crc kubenswrapper[4828]: E1210 20:38:49.722313 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"595a5aa7d2ce20a033806f8b57bc55abdcc874db342789ff09fac94113b9ab34\": container with ID starting with 595a5aa7d2ce20a033806f8b57bc55abdcc874db342789ff09fac94113b9ab34 not found: ID does not exist" containerID="595a5aa7d2ce20a033806f8b57bc55abdcc874db342789ff09fac94113b9ab34" Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.722348 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"595a5aa7d2ce20a033806f8b57bc55abdcc874db342789ff09fac94113b9ab34"} err="failed to get container status \"595a5aa7d2ce20a033806f8b57bc55abdcc874db342789ff09fac94113b9ab34\": rpc error: code = NotFound desc = could not find container \"595a5aa7d2ce20a033806f8b57bc55abdcc874db342789ff09fac94113b9ab34\": container with ID starting with 595a5aa7d2ce20a033806f8b57bc55abdcc874db342789ff09fac94113b9ab34 not found: ID does not exist" Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.722369 4828 scope.go:117] "RemoveContainer" containerID="e89bf57bf4777d6254ee68f5aacc94067c52005dbfca38f1318de3de79700b0a" Dec 10 20:38:49 crc kubenswrapper[4828]: E1210 20:38:49.723138 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e89bf57bf4777d6254ee68f5aacc94067c52005dbfca38f1318de3de79700b0a\": container with ID starting with e89bf57bf4777d6254ee68f5aacc94067c52005dbfca38f1318de3de79700b0a not found: ID does not exist" containerID="e89bf57bf4777d6254ee68f5aacc94067c52005dbfca38f1318de3de79700b0a" Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.723212 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e89bf57bf4777d6254ee68f5aacc94067c52005dbfca38f1318de3de79700b0a"} err="failed to get container status \"e89bf57bf4777d6254ee68f5aacc94067c52005dbfca38f1318de3de79700b0a\": rpc error: code = NotFound desc = could not find container \"e89bf57bf4777d6254ee68f5aacc94067c52005dbfca38f1318de3de79700b0a\": container with ID starting with e89bf57bf4777d6254ee68f5aacc94067c52005dbfca38f1318de3de79700b0a not found: ID does not exist" Dec 10 20:38:49 crc kubenswrapper[4828]: I1210 20:38:49.802510 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8" path="/var/lib/kubelet/pods/2b7f18a3-0eb6-414f-a95c-0ab0e9d36bb8/volumes" Dec 10 20:38:55 crc kubenswrapper[4828]: I1210 20:38:55.967786 4828 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-95vqh" Dec 10 20:38:56 crc kubenswrapper[4828]: I1210 20:38:56.032544 4828 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-95vqh" Dec 10 20:38:56 crc kubenswrapper[4828]: I1210 20:38:56.203558 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-95vqh"] Dec 10 20:38:57 crc kubenswrapper[4828]: I1210 20:38:57.690330 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-95vqh" podUID="9013fd72-548b-4f6b-9e6e-e0fdd9541900" containerName="registry-server" containerID="cri-o://4520def2eb44b069057a92cb95ae11e8936607be486b7a582559cadeaca841b5" gracePeriod=2 Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.266263 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-95vqh" Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.364693 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncj2p\" (UniqueName: \"kubernetes.io/projected/9013fd72-548b-4f6b-9e6e-e0fdd9541900-kube-api-access-ncj2p\") pod \"9013fd72-548b-4f6b-9e6e-e0fdd9541900\" (UID: \"9013fd72-548b-4f6b-9e6e-e0fdd9541900\") " Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.364871 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9013fd72-548b-4f6b-9e6e-e0fdd9541900-catalog-content\") pod \"9013fd72-548b-4f6b-9e6e-e0fdd9541900\" (UID: \"9013fd72-548b-4f6b-9e6e-e0fdd9541900\") " Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.365101 4828 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9013fd72-548b-4f6b-9e6e-e0fdd9541900-utilities\") pod \"9013fd72-548b-4f6b-9e6e-e0fdd9541900\" (UID: \"9013fd72-548b-4f6b-9e6e-e0fdd9541900\") " Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.366363 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9013fd72-548b-4f6b-9e6e-e0fdd9541900-utilities" (OuterVolumeSpecName: "utilities") pod "9013fd72-548b-4f6b-9e6e-e0fdd9541900" (UID: "9013fd72-548b-4f6b-9e6e-e0fdd9541900"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.375160 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9013fd72-548b-4f6b-9e6e-e0fdd9541900-kube-api-access-ncj2p" (OuterVolumeSpecName: "kube-api-access-ncj2p") pod "9013fd72-548b-4f6b-9e6e-e0fdd9541900" (UID: "9013fd72-548b-4f6b-9e6e-e0fdd9541900"). InnerVolumeSpecName "kube-api-access-ncj2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.468349 4828 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9013fd72-548b-4f6b-9e6e-e0fdd9541900-utilities\") on node \"crc\" DevicePath \"\"" Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.468434 4828 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncj2p\" (UniqueName: \"kubernetes.io/projected/9013fd72-548b-4f6b-9e6e-e0fdd9541900-kube-api-access-ncj2p\") on node \"crc\" DevicePath \"\"" Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.495329 4828 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9013fd72-548b-4f6b-9e6e-e0fdd9541900-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9013fd72-548b-4f6b-9e6e-e0fdd9541900" (UID: "9013fd72-548b-4f6b-9e6e-e0fdd9541900"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.572257 4828 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9013fd72-548b-4f6b-9e6e-e0fdd9541900-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.707738 4828 generic.go:334] "Generic (PLEG): container finished" podID="9013fd72-548b-4f6b-9e6e-e0fdd9541900" containerID="4520def2eb44b069057a92cb95ae11e8936607be486b7a582559cadeaca841b5" exitCode=0 Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.707790 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95vqh" event={"ID":"9013fd72-548b-4f6b-9e6e-e0fdd9541900","Type":"ContainerDied","Data":"4520def2eb44b069057a92cb95ae11e8936607be486b7a582559cadeaca841b5"} Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.707838 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-95vqh" event={"ID":"9013fd72-548b-4f6b-9e6e-e0fdd9541900","Type":"ContainerDied","Data":"066fcdf72d5cc0f655133d186e3f944c791067debd08b5fae07f74bf5cec886a"} Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.707855 4828 scope.go:117] "RemoveContainer" containerID="4520def2eb44b069057a92cb95ae11e8936607be486b7a582559cadeaca841b5" Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.708218 4828 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-95vqh" Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.735618 4828 scope.go:117] "RemoveContainer" containerID="6fb81de34afd424e41abe011dd5712c4ef67e8b901683b256536cb08998fe53a" Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.764059 4828 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-95vqh"] Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.770790 4828 scope.go:117] "RemoveContainer" containerID="6d701ad72875fbbeaddbed8af2aca59f216f62d35808423f7d44a61ebc21360d" Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.774627 4828 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-95vqh"] Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.821741 4828 scope.go:117] "RemoveContainer" containerID="4520def2eb44b069057a92cb95ae11e8936607be486b7a582559cadeaca841b5" Dec 10 20:38:58 crc kubenswrapper[4828]: E1210 20:38:58.822580 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4520def2eb44b069057a92cb95ae11e8936607be486b7a582559cadeaca841b5\": container with ID starting with 4520def2eb44b069057a92cb95ae11e8936607be486b7a582559cadeaca841b5 not found: ID does not exist" containerID="4520def2eb44b069057a92cb95ae11e8936607be486b7a582559cadeaca841b5" Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.822629 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4520def2eb44b069057a92cb95ae11e8936607be486b7a582559cadeaca841b5"} err="failed to get container status \"4520def2eb44b069057a92cb95ae11e8936607be486b7a582559cadeaca841b5\": rpc error: code = NotFound desc = could not find container \"4520def2eb44b069057a92cb95ae11e8936607be486b7a582559cadeaca841b5\": container with ID starting with 4520def2eb44b069057a92cb95ae11e8936607be486b7a582559cadeaca841b5 not found: ID does not exist" Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.822659 4828 scope.go:117] "RemoveContainer" containerID="6fb81de34afd424e41abe011dd5712c4ef67e8b901683b256536cb08998fe53a" Dec 10 20:38:58 crc kubenswrapper[4828]: E1210 20:38:58.823129 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fb81de34afd424e41abe011dd5712c4ef67e8b901683b256536cb08998fe53a\": container with ID starting with 6fb81de34afd424e41abe011dd5712c4ef67e8b901683b256536cb08998fe53a not found: ID does not exist" containerID="6fb81de34afd424e41abe011dd5712c4ef67e8b901683b256536cb08998fe53a" Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.823166 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fb81de34afd424e41abe011dd5712c4ef67e8b901683b256536cb08998fe53a"} err="failed to get container status \"6fb81de34afd424e41abe011dd5712c4ef67e8b901683b256536cb08998fe53a\": rpc error: code = NotFound desc = could not find container \"6fb81de34afd424e41abe011dd5712c4ef67e8b901683b256536cb08998fe53a\": container with ID starting with 6fb81de34afd424e41abe011dd5712c4ef67e8b901683b256536cb08998fe53a not found: ID does not exist" Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.823186 4828 scope.go:117] "RemoveContainer" containerID="6d701ad72875fbbeaddbed8af2aca59f216f62d35808423f7d44a61ebc21360d" Dec 10 20:38:58 crc kubenswrapper[4828]: E1210 20:38:58.824059 4828 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d701ad72875fbbeaddbed8af2aca59f216f62d35808423f7d44a61ebc21360d\": container with ID starting with 6d701ad72875fbbeaddbed8af2aca59f216f62d35808423f7d44a61ebc21360d not found: ID does not exist" containerID="6d701ad72875fbbeaddbed8af2aca59f216f62d35808423f7d44a61ebc21360d" Dec 10 20:38:58 crc kubenswrapper[4828]: I1210 20:38:58.824090 4828 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d701ad72875fbbeaddbed8af2aca59f216f62d35808423f7d44a61ebc21360d"} err="failed to get container status \"6d701ad72875fbbeaddbed8af2aca59f216f62d35808423f7d44a61ebc21360d\": rpc error: code = NotFound desc = could not find container \"6d701ad72875fbbeaddbed8af2aca59f216f62d35808423f7d44a61ebc21360d\": container with ID starting with 6d701ad72875fbbeaddbed8af2aca59f216f62d35808423f7d44a61ebc21360d not found: ID does not exist" Dec 10 20:38:59 crc kubenswrapper[4828]: I1210 20:38:59.804780 4828 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9013fd72-548b-4f6b-9e6e-e0fdd9541900" path="/var/lib/kubelet/pods/9013fd72-548b-4f6b-9e6e-e0fdd9541900/volumes" Dec 10 20:40:51 crc kubenswrapper[4828]: I1210 20:40:51.230774 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:40:51 crc kubenswrapper[4828]: I1210 20:40:51.232041 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:41:21 crc kubenswrapper[4828]: I1210 20:41:21.230451 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:41:21 crc kubenswrapper[4828]: I1210 20:41:21.231094 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:41:51 crc kubenswrapper[4828]: I1210 20:41:51.230029 4828 patch_prober.go:28] interesting pod/machine-config-daemon-677ql container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 10 20:41:51 crc kubenswrapper[4828]: I1210 20:41:51.230563 4828 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 10 20:41:51 crc kubenswrapper[4828]: I1210 20:41:51.230611 4828 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-677ql" Dec 10 20:41:51 crc kubenswrapper[4828]: I1210 20:41:51.231495 4828 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"039316a83fc8a6974ee10f5ecdceec1e0fbd066ac42e73d4bf48cef8444d9fe2"} pod="openshift-machine-config-operator/machine-config-daemon-677ql" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 10 20:41:51 crc kubenswrapper[4828]: I1210 20:41:51.231547 4828 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerName="machine-config-daemon" containerID="cri-o://039316a83fc8a6974ee10f5ecdceec1e0fbd066ac42e73d4bf48cef8444d9fe2" gracePeriod=600 Dec 10 20:41:51 crc kubenswrapper[4828]: E1210 20:41:51.366665 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:41:51 crc kubenswrapper[4828]: I1210 20:41:51.853097 4828 generic.go:334] "Generic (PLEG): container finished" podID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" containerID="039316a83fc8a6974ee10f5ecdceec1e0fbd066ac42e73d4bf48cef8444d9fe2" exitCode=0 Dec 10 20:41:51 crc kubenswrapper[4828]: I1210 20:41:51.853200 4828 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-677ql" event={"ID":"c0bc5145-c8c4-46a3-adf1-5d64ac422dc4","Type":"ContainerDied","Data":"039316a83fc8a6974ee10f5ecdceec1e0fbd066ac42e73d4bf48cef8444d9fe2"} Dec 10 20:41:51 crc kubenswrapper[4828]: I1210 20:41:51.853291 4828 scope.go:117] "RemoveContainer" containerID="325ae6fd60148173e9b100cf5cb87a1e61d379c9db65348aeb2ae47852a05290" Dec 10 20:41:51 crc kubenswrapper[4828]: I1210 20:41:51.854540 4828 scope.go:117] "RemoveContainer" containerID="039316a83fc8a6974ee10f5ecdceec1e0fbd066ac42e73d4bf48cef8444d9fe2" Dec 10 20:41:51 crc kubenswrapper[4828]: E1210 20:41:51.855339 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:42:05 crc kubenswrapper[4828]: I1210 20:42:05.789026 4828 scope.go:117] "RemoveContainer" containerID="039316a83fc8a6974ee10f5ecdceec1e0fbd066ac42e73d4bf48cef8444d9fe2" Dec 10 20:42:05 crc kubenswrapper[4828]: E1210 20:42:05.789783 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:42:20 crc kubenswrapper[4828]: I1210 20:42:20.789643 4828 scope.go:117] "RemoveContainer" containerID="039316a83fc8a6974ee10f5ecdceec1e0fbd066ac42e73d4bf48cef8444d9fe2" Dec 10 20:42:20 crc kubenswrapper[4828]: E1210 20:42:20.790505 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:42:32 crc kubenswrapper[4828]: I1210 20:42:32.789623 4828 scope.go:117] "RemoveContainer" containerID="039316a83fc8a6974ee10f5ecdceec1e0fbd066ac42e73d4bf48cef8444d9fe2" Dec 10 20:42:32 crc kubenswrapper[4828]: E1210 20:42:32.790711 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" Dec 10 20:42:44 crc kubenswrapper[4828]: I1210 20:42:44.788192 4828 scope.go:117] "RemoveContainer" containerID="039316a83fc8a6974ee10f5ecdceec1e0fbd066ac42e73d4bf48cef8444d9fe2" Dec 10 20:42:44 crc kubenswrapper[4828]: E1210 20:42:44.789036 4828 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-677ql_openshift-machine-config-operator(c0bc5145-c8c4-46a3-adf1-5d64ac422dc4)\"" pod="openshift-machine-config-operator/machine-config-daemon-677ql" podUID="c0bc5145-c8c4-46a3-adf1-5d64ac422dc4" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515116355527024457 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015116355527017374 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015116340516016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015116340516015457 5ustar corecore